var/home/core/zuul-output/0000755000175000017500000000000015114037120014517 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114057233015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005373706315114057223017713 0ustar rootrootDec 03 13:46:47 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 13:46:48 crc restorecon[4676]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 13:46:48 crc restorecon[4676]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 13:46:49 crc kubenswrapper[4677]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:46:49 crc kubenswrapper[4677]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 13:46:49 crc kubenswrapper[4677]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:46:49 crc kubenswrapper[4677]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:46:49 crc kubenswrapper[4677]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 13:46:49 crc kubenswrapper[4677]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.739937 4677 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746059 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746092 4677 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746098 4677 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746102 4677 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746106 4677 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746111 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746116 4677 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746121 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746126 4677 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746130 4677 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746134 4677 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746138 4677 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746142 4677 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746146 4677 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746150 4677 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746154 4677 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746158 4677 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746162 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746166 4677 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746170 4677 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746173 4677 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746177 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746181 4677 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746185 4677 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746189 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746193 4677 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746197 4677 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746201 4677 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746204 4677 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746208 4677 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746212 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746216 4677 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746220 4677 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746223 4677 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746227 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746239 4677 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746243 4677 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746247 4677 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746252 4677 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746257 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746261 4677 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746265 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746269 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746273 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746277 4677 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746280 4677 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746284 4677 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746288 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746292 4677 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746295 4677 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746299 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746303 4677 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746308 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746312 4677 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746317 4677 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746321 4677 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746325 4677 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746329 4677 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746334 4677 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746340 4677 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746345 4677 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746349 4677 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746354 4677 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746357 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746361 4677 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746365 4677 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746370 4677 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746373 4677 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746377 4677 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746381 4677 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.746384 4677 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746666 4677 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746680 4677 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746689 4677 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746695 4677 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746702 4677 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746706 4677 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746712 4677 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746718 4677 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746722 4677 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746727 4677 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746731 4677 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746736 4677 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746740 4677 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746744 4677 flags.go:64] FLAG: --cgroup-root="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746748 4677 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746753 4677 flags.go:64] FLAG: --client-ca-file="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746756 4677 flags.go:64] FLAG: --cloud-config="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746760 4677 flags.go:64] FLAG: --cloud-provider="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746765 4677 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746770 4677 flags.go:64] FLAG: --cluster-domain="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746774 4677 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746779 4677 flags.go:64] FLAG: --config-dir="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746783 4677 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746788 4677 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746794 4677 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746799 4677 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746803 4677 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746808 4677 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746813 4677 flags.go:64] FLAG: --contention-profiling="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746817 4677 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746822 4677 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746827 4677 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746831 4677 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746839 4677 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746843 4677 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746847 4677 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746851 4677 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746855 4677 flags.go:64] FLAG: --enable-server="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746859 4677 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746865 4677 flags.go:64] FLAG: --event-burst="100" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746869 4677 flags.go:64] FLAG: --event-qps="50" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746873 4677 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746877 4677 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746881 4677 flags.go:64] FLAG: --eviction-hard="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746887 4677 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746891 4677 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746895 4677 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746900 4677 flags.go:64] FLAG: --eviction-soft="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746904 4677 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746908 4677 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746912 4677 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746916 4677 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746920 4677 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746924 4677 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746928 4677 flags.go:64] FLAG: --feature-gates="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746933 4677 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746937 4677 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746941 4677 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746961 4677 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746965 4677 flags.go:64] FLAG: --healthz-port="10248" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746970 4677 flags.go:64] FLAG: --help="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746974 4677 flags.go:64] FLAG: --hostname-override="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746978 4677 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746982 4677 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746986 4677 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746991 4677 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.746995 4677 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747000 4677 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747004 4677 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747008 4677 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747013 4677 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747017 4677 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747022 4677 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747026 4677 flags.go:64] FLAG: --kube-reserved="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747030 4677 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747034 4677 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747038 4677 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747042 4677 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747047 4677 flags.go:64] FLAG: --lock-file="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747051 4677 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747055 4677 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747060 4677 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747066 4677 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747071 4677 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747076 4677 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747080 4677 flags.go:64] FLAG: --logging-format="text" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747085 4677 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747090 4677 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747094 4677 flags.go:64] FLAG: --manifest-url="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747099 4677 flags.go:64] FLAG: --manifest-url-header="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747105 4677 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747110 4677 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747115 4677 flags.go:64] FLAG: --max-pods="110" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747119 4677 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747124 4677 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747128 4677 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747132 4677 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747137 4677 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747141 4677 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747145 4677 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747156 4677 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747160 4677 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747164 4677 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747169 4677 flags.go:64] FLAG: --pod-cidr="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747172 4677 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747200 4677 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747204 4677 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747209 4677 flags.go:64] FLAG: --pods-per-core="0" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747213 4677 flags.go:64] FLAG: --port="10250" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747217 4677 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747221 4677 flags.go:64] FLAG: --provider-id="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747225 4677 flags.go:64] FLAG: --qos-reserved="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747229 4677 flags.go:64] FLAG: --read-only-port="10255" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747233 4677 flags.go:64] FLAG: --register-node="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747237 4677 flags.go:64] FLAG: --register-schedulable="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747241 4677 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747248 4677 flags.go:64] FLAG: --registry-burst="10" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747253 4677 flags.go:64] FLAG: --registry-qps="5" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747257 4677 flags.go:64] FLAG: --reserved-cpus="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747262 4677 flags.go:64] FLAG: --reserved-memory="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747268 4677 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747272 4677 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747277 4677 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747281 4677 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747285 4677 flags.go:64] FLAG: --runonce="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747290 4677 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747294 4677 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747299 4677 flags.go:64] FLAG: --seccomp-default="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747303 4677 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747307 4677 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747312 4677 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747316 4677 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747321 4677 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747324 4677 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747329 4677 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747332 4677 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747336 4677 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747341 4677 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747345 4677 flags.go:64] FLAG: --system-cgroups="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747349 4677 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747355 4677 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747359 4677 flags.go:64] FLAG: --tls-cert-file="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747363 4677 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747367 4677 flags.go:64] FLAG: --tls-min-version="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747372 4677 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747376 4677 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747380 4677 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747384 4677 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747388 4677 flags.go:64] FLAG: --v="2" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747395 4677 flags.go:64] FLAG: --version="false" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747401 4677 flags.go:64] FLAG: --vmodule="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747406 4677 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747410 4677 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747551 4677 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747556 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747561 4677 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747565 4677 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747568 4677 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747572 4677 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747576 4677 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747580 4677 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747583 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747587 4677 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747591 4677 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747594 4677 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747598 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747601 4677 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747605 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747608 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747612 4677 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747615 4677 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747618 4677 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747622 4677 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747625 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747629 4677 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747634 4677 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747638 4677 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747641 4677 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747645 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747648 4677 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747652 4677 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747659 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747663 4677 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747666 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747670 4677 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747673 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747677 4677 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747681 4677 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747684 4677 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747687 4677 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747691 4677 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747695 4677 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747699 4677 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747702 4677 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747707 4677 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747711 4677 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747714 4677 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747718 4677 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747721 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747725 4677 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747728 4677 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747732 4677 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747735 4677 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747738 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747742 4677 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747745 4677 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747750 4677 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747753 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747759 4677 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747763 4677 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747768 4677 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747773 4677 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747777 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747783 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747787 4677 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747792 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747796 4677 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747799 4677 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747803 4677 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747806 4677 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747810 4677 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747813 4677 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747817 4677 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.747820 4677 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.747830 4677 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.756279 4677 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.756330 4677 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756459 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756476 4677 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756484 4677 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756491 4677 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756496 4677 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756502 4677 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756507 4677 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756512 4677 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756517 4677 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756523 4677 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756528 4677 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756533 4677 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756538 4677 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756543 4677 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756548 4677 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756553 4677 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756558 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756563 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756568 4677 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756573 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756578 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756582 4677 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756587 4677 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756592 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756598 4677 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756603 4677 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756609 4677 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756617 4677 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756623 4677 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756629 4677 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756634 4677 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756640 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756644 4677 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756649 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756656 4677 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756661 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756666 4677 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756671 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756676 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756681 4677 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756686 4677 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756691 4677 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756697 4677 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756704 4677 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756709 4677 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756714 4677 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756718 4677 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756724 4677 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756728 4677 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756735 4677 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756741 4677 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756746 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756752 4677 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756758 4677 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756763 4677 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756770 4677 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756776 4677 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756781 4677 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756786 4677 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756791 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756796 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756801 4677 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756806 4677 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756811 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756816 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756821 4677 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756826 4677 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756831 4677 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756837 4677 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756841 4677 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.756852 4677 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.756861 4677 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757053 4677 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757068 4677 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757074 4677 feature_gate.go:330] unrecognized feature gate: Example Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757080 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757085 4677 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757091 4677 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757095 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757100 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757106 4677 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757111 4677 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757116 4677 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757121 4677 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757126 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757131 4677 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757136 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757141 4677 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757145 4677 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757150 4677 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757155 4677 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757160 4677 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757165 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757169 4677 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757174 4677 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757179 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757185 4677 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757190 4677 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757194 4677 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757199 4677 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757204 4677 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757211 4677 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757219 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757224 4677 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757230 4677 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757236 4677 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757243 4677 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757248 4677 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757255 4677 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757261 4677 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757267 4677 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757272 4677 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757277 4677 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757282 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757287 4677 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757292 4677 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757297 4677 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757302 4677 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757307 4677 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757312 4677 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757316 4677 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757321 4677 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757326 4677 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757331 4677 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757338 4677 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757344 4677 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757349 4677 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757357 4677 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757365 4677 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757378 4677 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757388 4677 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757395 4677 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757401 4677 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757408 4677 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757415 4677 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757422 4677 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757430 4677 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757438 4677 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757445 4677 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757451 4677 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757457 4677 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757464 4677 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.757473 4677 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.757483 4677 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.757718 4677 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.761465 4677 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.761569 4677 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.762208 4677 server.go:997] "Starting client certificate rotation" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.762242 4677 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.762705 4677 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-22 12:14:47.333004732 +0000 UTC Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.762841 4677 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.769340 4677 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.771071 4677 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.771716 4677 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.803366 4677 log.go:25] "Validated CRI v1 runtime API" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.818927 4677 log.go:25] "Validated CRI v1 image API" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.820244 4677 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.824030 4677 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-13-41-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.824059 4677 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.842918 4677 manager.go:217] Machine: {Timestamp:2025-12-03 13:46:49.83984422 +0000 UTC m=+0.586176715 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1bc9448c-c189-4fc5-b5df-6e0a45902353 BootID:f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b5:a7:16 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b5:a7:16 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:eb:28:66 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e8:98:1a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:99:a5:e2 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a6:52:18 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8a:2a:dc:49:24:09 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fa:b3:a6:0a:f6:79 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.843245 4677 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.843392 4677 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.845716 4677 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.845927 4677 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.845990 4677 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.846241 4677 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.846254 4677 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.846475 4677 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.846511 4677 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.846696 4677 state_mem.go:36] "Initialized new in-memory state store" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.846805 4677 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.847778 4677 kubelet.go:418] "Attempting to sync node with API server" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.847803 4677 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.847822 4677 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.847836 4677 kubelet.go:324] "Adding apiserver pod source" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.847850 4677 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.874818 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.874924 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.874178 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.874978 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.877414 4677 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.877828 4677 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.878581 4677 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879247 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879281 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879292 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879300 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879314 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879324 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879333 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879346 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879358 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879367 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879380 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879389 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.879754 4677 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.880269 4677 server.go:1280] "Started kubelet" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.881736 4677 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.881753 4677 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.882266 4677 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 13:46:49 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.883781 4677 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.883825 4677 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.883967 4677 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.883996 4677 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 00:10:45.499432263 +0000 UTC Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.884056 4677 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 442h23m55.615379236s for next certificate rotation Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.884360 4677 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.884370 4677 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.884438 4677 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885026 4677 server.go:460] "Adding debug handlers to kubelet server" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885043 4677 factory.go:55] Registering systemd factory Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885064 4677 factory.go:221] Registration of the systemd container factory successfully Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885363 4677 factory.go:153] Registering CRI-O factory Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885393 4677 factory.go:221] Registration of the crio container factory successfully Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885493 4677 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885519 4677 factory.go:103] Registering Raw factory Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.885538 4677 manager.go:1196] Started watching for new ooms in manager Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.885551 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.885622 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.886409 4677 manager.go:319] Starting recovery of all containers Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.886384 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="200ms" Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.885766 4677 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.162:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187db89c2b50f36f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:46:49.880236911 +0000 UTC m=+0.626569376,LastTimestamp:2025-12-03 13:46:49.880236911 +0000 UTC m=+0.626569376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891337 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891393 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891412 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891426 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891438 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891451 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891461 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891474 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891486 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891496 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891507 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891517 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891528 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891539 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891550 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891563 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891604 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891636 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891649 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891661 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891674 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891685 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891696 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891708 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891718 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891731 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891746 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891757 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891768 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891777 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891788 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891799 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891826 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891836 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891846 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891855 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891866 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891877 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891888 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891900 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891914 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891923 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891934 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891962 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891975 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891984 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.891994 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892006 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892020 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892032 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892043 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892055 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892070 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892082 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892093 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892104 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892117 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892128 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892139 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892149 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892161 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892172 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892183 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892193 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892205 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892214 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892224 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892233 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892243 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892254 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892264 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892276 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892294 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892304 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892315 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892327 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892339 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892351 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892362 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892376 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892389 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892402 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892417 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892428 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892439 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892451 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892462 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892473 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892484 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892495 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892506 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892516 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892528 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892543 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892552 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892562 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892574 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892583 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892592 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892603 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892611 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892620 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892633 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892647 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892696 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892711 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892724 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892737 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892749 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892762 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892775 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892788 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892802 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892813 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892825 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892836 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892848 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892860 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892871 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892881 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892890 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892901 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892910 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892919 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892929 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892938 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892965 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892974 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.892988 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893000 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893011 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893025 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893038 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893048 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893057 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893070 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893080 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893092 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893102 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893111 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893121 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893130 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893140 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893151 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893168 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893179 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893192 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893202 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893211 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893221 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893232 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893242 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893253 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893264 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893275 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893286 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893295 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893310 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893321 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893333 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893346 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893357 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893370 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893385 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893398 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893409 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893421 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893433 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893445 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893456 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893467 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893477 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.893487 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927159 4677 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927435 4677 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927515 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927555 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927576 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927593 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927622 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927639 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927661 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927675 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927688 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927702 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927719 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927735 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927750 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927765 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927784 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927799 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927818 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927835 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927850 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927867 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927884 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927902 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927916 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927929 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927970 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.927987 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928008 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928021 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928034 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928051 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928064 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928084 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928097 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928110 4677 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928124 4677 reconstruct.go:97] "Volume reconstruction finished" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.928139 4677 reconciler.go:26] "Reconciler: start to sync state" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.938707 4677 manager.go:324] Recovery completed Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.953206 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.957927 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.957984 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.957995 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.959320 4677 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.959337 4677 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.959359 4677 state_mem.go:36] "Initialized new in-memory state store" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.970785 4677 policy_none.go:49] "None policy: Start" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.972008 4677 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.972037 4677 state_mem.go:35] "Initializing new in-memory state store" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.973080 4677 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.974707 4677 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.974755 4677 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 13:46:49 crc kubenswrapper[4677]: I1203 13:46:49.974785 4677 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.974828 4677 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 13:46:49 crc kubenswrapper[4677]: W1203 13:46:49.976849 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.976933 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:49 crc kubenswrapper[4677]: E1203 13:46:49.984330 4677 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.043228 4677 manager.go:334] "Starting Device Plugin manager" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.043286 4677 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.043299 4677 server.go:79] "Starting device plugin registration server" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.043740 4677 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.043753 4677 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.044200 4677 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.044272 4677 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.044279 4677 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 13:46:50 crc kubenswrapper[4677]: E1203 13:46:50.050324 4677 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.075927 4677 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.076076 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.077244 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.077283 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.077295 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.077414 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.077730 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.077785 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078300 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078333 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078346 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078464 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078570 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078601 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078966 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078988 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.078999 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079410 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079469 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079482 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079519 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079552 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079562 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079689 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079797 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.079858 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.080695 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.080733 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.080745 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.080900 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.080900 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.080981 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.080991 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081001 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081036 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081748 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081783 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081752 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081795 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081812 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.081836 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.082033 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.082548 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.083165 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.083195 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.083207 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: E1203 13:46:50.087208 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="400ms" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132240 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132280 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132301 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132321 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132341 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132358 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132375 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132427 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132469 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132493 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132515 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132540 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132564 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132584 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.132603 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.144037 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.145445 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.145502 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.145516 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.145549 4677 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:46:50 crc kubenswrapper[4677]: E1203 13:46:50.146148 4677 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.233702 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.233798 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.233843 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.233903 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.233967 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234025 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234086 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234028 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234058 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234151 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234177 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234261 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234216 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234389 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234445 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234466 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234502 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234236 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234539 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234561 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234568 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234587 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234598 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234616 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234662 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234684 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234689 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234749 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234791 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.234891 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.346283 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.347574 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.347615 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.347624 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.347653 4677 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:46:50 crc kubenswrapper[4677]: E1203 13:46:50.348176 4677 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.421789 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.429850 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.448754 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.462456 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: W1203 13:46:50.466085 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a7295df14d086189a215b6e4aca965cfc1c81c904076fbbbc8d65a718db46922 WatchSource:0}: Error finding container a7295df14d086189a215b6e4aca965cfc1c81c904076fbbbc8d65a718db46922: Status 404 returned error can't find the container with id a7295df14d086189a215b6e4aca965cfc1c81c904076fbbbc8d65a718db46922 Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.466477 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:50 crc kubenswrapper[4677]: W1203 13:46:50.469578 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-75ed0149cdc16711c2615c4dd7084d87cc066d48f9e67d80bd1ae339d07719ae WatchSource:0}: Error finding container 75ed0149cdc16711c2615c4dd7084d87cc066d48f9e67d80bd1ae339d07719ae: Status 404 returned error can't find the container with id 75ed0149cdc16711c2615c4dd7084d87cc066d48f9e67d80bd1ae339d07719ae Dec 03 13:46:50 crc kubenswrapper[4677]: W1203 13:46:50.471593 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-dd295efabb1aced7c89f9810c78d667655ab213d76df3f06ebf456e96bcef6d1 WatchSource:0}: Error finding container dd295efabb1aced7c89f9810c78d667655ab213d76df3f06ebf456e96bcef6d1: Status 404 returned error can't find the container with id dd295efabb1aced7c89f9810c78d667655ab213d76df3f06ebf456e96bcef6d1 Dec 03 13:46:50 crc kubenswrapper[4677]: W1203 13:46:50.480560 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-37e54128bad6464ac735cc93302414bdc629ee7a87495890c9e42e3ec8c8fb35 WatchSource:0}: Error finding container 37e54128bad6464ac735cc93302414bdc629ee7a87495890c9e42e3ec8c8fb35: Status 404 returned error can't find the container with id 37e54128bad6464ac735cc93302414bdc629ee7a87495890c9e42e3ec8c8fb35 Dec 03 13:46:50 crc kubenswrapper[4677]: E1203 13:46:50.488431 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="800ms" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.748547 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.750110 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.750155 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.750165 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.750192 4677 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:46:50 crc kubenswrapper[4677]: E1203 13:46:50.750638 4677 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Dec 03 13:46:50 crc kubenswrapper[4677]: W1203 13:46:50.861497 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:50 crc kubenswrapper[4677]: E1203 13:46:50.861586 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.928287 4677 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.980906 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.981125 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"37e54128bad6464ac735cc93302414bdc629ee7a87495890c9e42e3ec8c8fb35"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.983416 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.983453 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dd295efabb1aced7c89f9810c78d667655ab213d76df3f06ebf456e96bcef6d1"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.983601 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.985346 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.985415 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.985428 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.985906 4677 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8" exitCode=0 Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.986017 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.986099 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"75ed0149cdc16711c2615c4dd7084d87cc066d48f9e67d80bd1ae339d07719ae"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.986441 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.987568 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.987606 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.987620 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.987884 4677 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570" exitCode=0 Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.987978 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.988003 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a7295df14d086189a215b6e4aca965cfc1c81c904076fbbbc8d65a718db46922"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.988076 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.988527 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.988727 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.988753 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.988765 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.989514 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.989542 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.989553 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.989736 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.989765 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fedd4fd6080e6101e0f82ffffc2abcf45f1afe6e682a491f8bbdf57eee35a766"} Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.989842 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.990517 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.990544 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:50 crc kubenswrapper[4677]: I1203 13:46:50.990555 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:51 crc kubenswrapper[4677]: W1203 13:46:51.137685 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:51 crc kubenswrapper[4677]: E1203 13:46:51.137777 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:51 crc kubenswrapper[4677]: E1203 13:46:51.289297 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="1.6s" Dec 03 13:46:51 crc kubenswrapper[4677]: W1203 13:46:51.314498 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:51 crc kubenswrapper[4677]: E1203 13:46:51.314590 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:51 crc kubenswrapper[4677]: I1203 13:46:51.551356 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:51 crc kubenswrapper[4677]: I1203 13:46:51.552794 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:51 crc kubenswrapper[4677]: I1203 13:46:51.552841 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:51 crc kubenswrapper[4677]: I1203 13:46:51.552857 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:51 crc kubenswrapper[4677]: I1203 13:46:51.552893 4677 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:46:51 crc kubenswrapper[4677]: E1203 13:46:51.553492 4677 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Dec 03 13:46:51 crc kubenswrapper[4677]: W1203 13:46:51.571828 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:51 crc kubenswrapper[4677]: E1203 13:46:51.571975 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:51 crc kubenswrapper[4677]: I1203 13:46:51.892481 4677 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 13:46:51 crc kubenswrapper[4677]: E1203 13:46:51.893757 4677 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Dec 03 13:46:51 crc kubenswrapper[4677]: I1203 13:46:51.928659 4677 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.001977 4677 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618" exitCode=0 Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.002055 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.002113 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.002128 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.002141 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.002266 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.006460 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.006501 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.006513 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.006598 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.007494 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.007518 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.007526 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.026726 4677 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3" exitCode=0 Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.026795 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.026852 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.026865 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.026877 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.028419 4677 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4" exitCode=0 Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.028490 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.028625 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.029414 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.029438 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.029449 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.032183 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d"} Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.032279 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.004561 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.034159 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.034173 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.035462 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.035504 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:52 crc kubenswrapper[4677]: I1203 13:46:52.035514 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:52 crc kubenswrapper[4677]: E1203 13:46:52.890555 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="3.2s" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.036216 4677 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb" exitCode=0 Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.036283 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb"} Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.036409 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.037348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.037371 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.037382 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.041284 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.041675 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.042344 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f"} Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.042479 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04"} Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.046092 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.046143 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.046157 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.056224 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.056671 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.056686 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.154151 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.155340 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.155374 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.155415 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.155443 4677 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.525214 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:53 crc kubenswrapper[4677]: I1203 13:46:53.753860 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.047764 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8"} Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.047814 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6"} Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.047824 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.047830 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c"} Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.047844 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26"} Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.047858 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.047933 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.048718 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.048750 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.048760 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.049710 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.049754 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.049770 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.162870 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.478681 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.479167 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.493628 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.493676 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:54 crc kubenswrapper[4677]: I1203 13:46:54.493684 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.057005 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412"} Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.057072 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.057246 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.057260 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.057703 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.058356 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.058398 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.058411 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.058529 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.058585 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.058603 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.059238 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.059281 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.059296 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:55 crc kubenswrapper[4677]: I1203 13:46:55.933135 4677 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 13:46:56 crc kubenswrapper[4677]: I1203 13:46:56.059675 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:56 crc kubenswrapper[4677]: I1203 13:46:56.060704 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:56 crc kubenswrapper[4677]: I1203 13:46:56.060786 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:56 crc kubenswrapper[4677]: I1203 13:46:56.060804 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:56 crc kubenswrapper[4677]: I1203 13:46:56.526092 4677 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 13:46:56 crc kubenswrapper[4677]: I1203 13:46:56.526163 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 13:46:57 crc kubenswrapper[4677]: I1203 13:46:57.228815 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:57 crc kubenswrapper[4677]: I1203 13:46:57.228993 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:46:57 crc kubenswrapper[4677]: I1203 13:46:57.229042 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:57 crc kubenswrapper[4677]: I1203 13:46:57.230288 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:57 crc kubenswrapper[4677]: I1203 13:46:57.230328 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:57 crc kubenswrapper[4677]: I1203 13:46:57.230341 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:57 crc kubenswrapper[4677]: I1203 13:46:57.392497 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.064402 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.065309 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.065354 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.065371 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.606931 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.607192 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.609156 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.609216 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:58 crc kubenswrapper[4677]: I1203 13:46:58.609233 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:59 crc kubenswrapper[4677]: I1203 13:46:59.713060 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:46:59 crc kubenswrapper[4677]: I1203 13:46:59.713262 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:46:59 crc kubenswrapper[4677]: I1203 13:46:59.714831 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:46:59 crc kubenswrapper[4677]: I1203 13:46:59.714877 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:46:59 crc kubenswrapper[4677]: I1203 13:46:59.714886 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:46:59 crc kubenswrapper[4677]: I1203 13:46:59.718501 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:47:00 crc kubenswrapper[4677]: E1203 13:47:00.050472 4677 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 13:47:00 crc kubenswrapper[4677]: I1203 13:47:00.070482 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:47:00 crc kubenswrapper[4677]: I1203 13:47:00.070762 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:47:00 crc kubenswrapper[4677]: I1203 13:47:00.071548 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:00 crc kubenswrapper[4677]: I1203 13:47:00.071582 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:00 crc kubenswrapper[4677]: I1203 13:47:00.071594 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:00 crc kubenswrapper[4677]: I1203 13:47:00.075814 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:47:01 crc kubenswrapper[4677]: I1203 13:47:01.073105 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:47:01 crc kubenswrapper[4677]: I1203 13:47:01.074217 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:01 crc kubenswrapper[4677]: I1203 13:47:01.074315 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:01 crc kubenswrapper[4677]: I1203 13:47:01.074331 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.075498 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.076663 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.076728 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.076749 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.233566 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.233794 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.235068 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.235140 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.235154 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:02 crc kubenswrapper[4677]: I1203 13:47:02.929296 4677 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 13:47:03 crc kubenswrapper[4677]: E1203 13:47:03.156506 4677 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 03 13:47:03 crc kubenswrapper[4677]: W1203 13:47:03.349186 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.349295 4677 trace.go:236] Trace[331235704]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:46:53.347) (total time: 10001ms): Dec 03 13:47:03 crc kubenswrapper[4677]: Trace[331235704]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:47:03.349) Dec 03 13:47:03 crc kubenswrapper[4677]: Trace[331235704]: [10.001455163s] [10.001455163s] END Dec 03 13:47:03 crc kubenswrapper[4677]: E1203 13:47:03.349324 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 13:47:03 crc kubenswrapper[4677]: W1203 13:47:03.583665 4677 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.583823 4677 trace.go:236] Trace[365144137]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:46:53.581) (total time: 10002ms): Dec 03 13:47:03 crc kubenswrapper[4677]: Trace[365144137]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:47:03.583) Dec 03 13:47:03 crc kubenswrapper[4677]: Trace[365144137]: [10.002110112s] [10.002110112s] END Dec 03 13:47:03 crc kubenswrapper[4677]: E1203 13:47:03.583857 4677 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.679768 4677 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.679895 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.687464 4677 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.687544 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.761265 4677 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Dec 03 13:47:03 crc kubenswrapper[4677]: I1203 13:47:03.761338 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 13:47:06 crc kubenswrapper[4677]: I1203 13:47:06.357249 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:47:06 crc kubenswrapper[4677]: I1203 13:47:06.358548 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:06 crc kubenswrapper[4677]: I1203 13:47:06.358586 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:06 crc kubenswrapper[4677]: I1203 13:47:06.358595 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:06 crc kubenswrapper[4677]: I1203 13:47:06.358616 4677 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:47:06 crc kubenswrapper[4677]: E1203 13:47:06.363105 4677 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 13:47:06 crc kubenswrapper[4677]: I1203 13:47:06.526478 4677 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 13:47:06 crc kubenswrapper[4677]: I1203 13:47:06.526595 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 13:47:07 crc kubenswrapper[4677]: I1203 13:47:07.166582 4677 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 13:47:08 crc kubenswrapper[4677]: E1203 13:47:08.675340 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.677923 4677 trace.go:236] Trace[1390200761]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:46:53.912) (total time: 14765ms): Dec 03 13:47:08 crc kubenswrapper[4677]: Trace[1390200761]: ---"Objects listed" error: 14765ms (13:47:08.677) Dec 03 13:47:08 crc kubenswrapper[4677]: Trace[1390200761]: [14.765599957s] [14.765599957s] END Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.678014 4677 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.678276 4677 trace.go:236] Trace[121317844]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 13:46:54.665) (total time: 14012ms): Dec 03 13:47:08 crc kubenswrapper[4677]: Trace[121317844]: ---"Objects listed" error: 14012ms (13:47:08.678) Dec 03 13:47:08 crc kubenswrapper[4677]: Trace[121317844]: [14.012645684s] [14.012645684s] END Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.678299 4677 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.678800 4677 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.689733 4677 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.727707 4677 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42380->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.727746 4677 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42388->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.727771 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42380->192.168.126.11:17697: read: connection reset by peer" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.727810 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42388->192.168.126.11:17697: read: connection reset by peer" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.760974 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.761773 4677 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.761908 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.766916 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.860661 4677 apiserver.go:52] "Watching apiserver" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.862445 4677 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.862693 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.863025 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.863049 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:08 crc kubenswrapper[4677]: E1203 13:47:08.863078 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:08 crc kubenswrapper[4677]: E1203 13:47:08.863114 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.863135 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.863252 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.863503 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.863548 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:08 crc kubenswrapper[4677]: E1203 13:47:08.863638 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866062 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866062 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866191 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866319 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866325 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866643 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866665 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.866785 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.867534 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.884933 4677 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.890291 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.900901 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.911929 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.923587 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.941369 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.951367 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.959238 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.970237 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.978648 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.979902 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.979968 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.979998 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980023 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980049 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980074 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980097 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980120 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980145 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980168 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980201 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980193 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980255 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980281 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980305 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980329 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980349 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980353 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980370 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980400 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980373 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980455 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980456 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980471 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980479 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980535 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980554 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980568 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980583 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980597 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980610 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980636 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980660 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980714 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980716 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980741 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980753 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980767 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980792 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980815 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980839 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980868 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980880 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980896 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980903 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.980927 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981027 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981036 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981063 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981087 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981110 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981132 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981191 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981215 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981271 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981300 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981357 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981382 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981402 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981424 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981447 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981472 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981494 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981130 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981517 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981541 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981569 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981588 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981614 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981640 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981743 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981772 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981797 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981859 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981882 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981904 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981926 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982002 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982031 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982088 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982109 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982135 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982158 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982182 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982204 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982224 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982246 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982269 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982297 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982317 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981146 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982331 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982337 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982363 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982388 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982412 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982434 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982457 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982477 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982496 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982516 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982535 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982561 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982584 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982610 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982637 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982660 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982682 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981143 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981190 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981210 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981239 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981244 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981312 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981355 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981394 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981410 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981432 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981483 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981489 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981580 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981592 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981621 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981672 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981690 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981723 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981737 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981794 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981820 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981853 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981896 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.981906 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982042 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982057 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982081 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982133 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982309 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982310 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982352 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982490 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982533 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982573 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982942 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.982931 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.983172 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.983453 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.983547 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.983621 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: E1203 13:47:08.983658 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:47:09.482688877 +0000 UTC m=+20.229021332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985077 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985110 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985139 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985164 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985188 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985219 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985241 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985259 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985281 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985306 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985329 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985352 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985376 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985398 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985423 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985449 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985471 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985492 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985512 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985528 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985544 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985559 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985578 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985594 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985616 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985636 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985658 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.985713 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.987625 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.983708 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.989296 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992058 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992090 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992113 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992129 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992147 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992162 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992178 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992194 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992209 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992224 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992240 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992221 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992256 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992359 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992376 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992383 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992415 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992444 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992476 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992505 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992540 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992580 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992603 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992622 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992641 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992670 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992704 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992795 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992829 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992852 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.992872 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.993477 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.993739 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994472 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994512 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994537 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994559 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994578 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994660 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994683 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994701 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994720 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994782 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994772 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994803 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994823 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994842 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994870 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994894 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994916 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994958 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.994942 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995006 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995029 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995052 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995073 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995095 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995117 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995138 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995160 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995183 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995207 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995229 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995254 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995278 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995305 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995331 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995356 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995382 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995404 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995427 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995455 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995474 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995490 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995506 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995523 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995539 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995555 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995568 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995572 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995635 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995662 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995686 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995780 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995816 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995837 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995859 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995880 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995896 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995913 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995936 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995976 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.995996 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996015 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996035 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996036 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996056 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996077 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996091 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996154 4677 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996169 4677 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996179 4677 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996191 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996201 4677 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996211 4677 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996222 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996207 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996232 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996300 4677 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996304 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996314 4677 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996328 4677 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996342 4677 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996352 4677 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996364 4677 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996374 4677 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996384 4677 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996398 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996410 4677 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996420 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996429 4677 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996580 4677 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996593 4677 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996602 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996612 4677 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996623 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996633 4677 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996642 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996652 4677 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996660 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996671 4677 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996680 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996690 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996700 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996715 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996725 4677 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996734 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996745 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996755 4677 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996770 4677 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996788 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996801 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996812 4677 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996823 4677 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996834 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996846 4677 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996859 4677 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996870 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996884 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996896 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996912 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996925 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996935 4677 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996966 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996979 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.996991 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997002 4677 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997014 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997014 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997059 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997061 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997072 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997083 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997097 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997110 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997121 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997132 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997142 4677 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997153 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997163 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997424 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.997419 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.998178 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.998572 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.998778 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:08 crc kubenswrapper[4677]: I1203 13:47:08.999357 4677 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.001607 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.001673 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.001847 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.002036 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.002038 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.002214 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.002450 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.002494 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.003001 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.003133 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.003285 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.003377 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.003690 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.003886 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.004027 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.004315 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:09.504291364 +0000 UTC m=+20.250623819 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.004390 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.004399 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.004750 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.004795 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.005024 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.005148 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.005442 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.005516 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.005738 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.005835 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.005871 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.006429 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.006692 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.007940 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.008116 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.008340 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.008440 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.009006 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.009312 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.010157 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.010620 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.010649 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.011031 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:09.510997402 +0000 UTC m=+20.257330087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.011410 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.011689 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.011844 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.012176 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.012277 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.012871 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.013660 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.014489 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.016266 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.016372 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.016474 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.016495 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.016508 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.016565 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:09.516547445 +0000 UTC m=+20.262879900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.016476 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.017288 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.017571 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.017661 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.017888 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.018823 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.020328 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.020679 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.021084 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.021264 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.021627 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.021642 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.021928 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.021985 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.022289 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.022764 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.024607 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.024666 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.024682 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.024770 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:09.524747246 +0000 UTC m=+20.271079701 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.028154 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.029068 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.029544 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.029880 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.031133 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.032066 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.032474 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.032747 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.033183 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.033306 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.033764 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.034190 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.034458 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.036288 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.036571 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.036739 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.037411 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.038093 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.038679 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.041113 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.042245 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.043233 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.043807 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.044198 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.044329 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.045386 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.045434 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.045461 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.045512 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.045903 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.047350 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.047464 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.047191 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050076 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050119 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050105 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050230 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050247 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050520 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050569 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.050910 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.051087 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.051364 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.051493 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.051909 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.052166 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.052205 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.052323 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.052676 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.054132 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.054779 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.054855 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.056719 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.056744 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.057068 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.057383 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.057458 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.057748 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.059478 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.059645 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.077479 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.082588 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.088536 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.097611 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.097647 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.097707 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.097718 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.100876 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.100905 4677 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.097800 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.100917 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.100928 4677 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.100940 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.098191 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.100967 4677 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101038 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101051 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.100774 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101063 4677 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101113 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101144 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101153 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101160 4677 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101168 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101178 4677 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101365 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101384 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101393 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101401 4677 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101410 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101419 4677 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101447 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101458 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101467 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101476 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101486 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101496 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101523 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101531 4677 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101540 4677 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101549 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101557 4677 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101566 4677 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101576 4677 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101603 4677 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101613 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101621 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101630 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101639 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101647 4677 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101655 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101683 4677 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101692 4677 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101701 4677 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101709 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101718 4677 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101728 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101753 4677 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101761 4677 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101769 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101778 4677 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101787 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101795 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101803 4677 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101811 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101819 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101827 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101849 4677 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101858 4677 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101866 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101874 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101883 4677 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101891 4677 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101899 4677 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101908 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101917 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101925 4677 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101932 4677 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101940 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101983 4677 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.101992 4677 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102000 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102008 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102017 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102026 4677 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102035 4677 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102043 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102052 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102060 4677 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102067 4677 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102075 4677 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102083 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102093 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102101 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102111 4677 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102142 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102151 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102160 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102168 4677 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102196 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102232 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102242 4677 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102252 4677 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102262 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102290 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102300 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102309 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102318 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102326 4677 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102359 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102367 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102375 4677 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102383 4677 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102391 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102400 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102409 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102417 4677 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102461 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102470 4677 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102478 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102486 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102494 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102503 4677 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102510 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102518 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102526 4677 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102534 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102542 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102550 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102752 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102766 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102778 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102788 4677 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102796 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.102805 4677 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.103911 4677 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f" exitCode=255 Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.103978 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f"} Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.109934 4677 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.110143 4677 scope.go:117] "RemoveContainer" containerID="888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.111859 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.122757 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.133521 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.146187 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.156497 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.167151 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.176052 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.177347 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.181688 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.187852 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 13:47:09 crc kubenswrapper[4677]: W1203 13:47:09.188838 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-d46fa10305ae36d7eb191372b80e0f3c2dbed0a5abf8da2fca12df53f7e1b262 WatchSource:0}: Error finding container d46fa10305ae36d7eb191372b80e0f3c2dbed0a5abf8da2fca12df53f7e1b262: Status 404 returned error can't find the container with id d46fa10305ae36d7eb191372b80e0f3c2dbed0a5abf8da2fca12df53f7e1b262 Dec 03 13:47:09 crc kubenswrapper[4677]: W1203 13:47:09.194607 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-ca49f4345a8dd9a2289278beb257ee8a8ccc3a68bf877395666637f0ebd6a520 WatchSource:0}: Error finding container ca49f4345a8dd9a2289278beb257ee8a8ccc3a68bf877395666637f0ebd6a520: Status 404 returned error can't find the container with id ca49f4345a8dd9a2289278beb257ee8a8ccc3a68bf877395666637f0ebd6a520 Dec 03 13:47:09 crc kubenswrapper[4677]: W1203 13:47:09.201528 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-853e1a511b94837123c4e1f9a8bc46b089a55efc7b1447e96ed9051fcefa14bf WatchSource:0}: Error finding container 853e1a511b94837123c4e1f9a8bc46b089a55efc7b1447e96ed9051fcefa14bf: Status 404 returned error can't find the container with id 853e1a511b94837123c4e1f9a8bc46b089a55efc7b1447e96ed9051fcefa14bf Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.505157 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.505241 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.505361 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.505415 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:47:10.505363287 +0000 UTC m=+21.251695742 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.505468 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:10.5054599 +0000 UTC m=+21.251792355 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.606552 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.606613 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.606634 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606771 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606805 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606847 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606857 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:10.606837014 +0000 UTC m=+21.353169469 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606861 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606920 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:10.606909946 +0000 UTC m=+21.353242411 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606806 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606978 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.606995 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.607038 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:10.60702895 +0000 UTC m=+21.353361405 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.684894 4677 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.975733 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:09 crc kubenswrapper[4677]: E1203 13:47:09.975851 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.983729 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.984674 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.986102 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.986896 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.988090 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.988646 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.989311 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.990297 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.990878 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.991877 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.992435 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.993255 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:09Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.993591 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.994075 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.994598 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.995610 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.996620 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.997572 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.997993 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.998537 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 13:47:09 crc kubenswrapper[4677]: I1203 13:47:09.999613 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.000156 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.001262 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.001763 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.002873 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.003505 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.004100 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.005372 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.005875 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.006878 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.007626 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.008595 4677 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.008731 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.011226 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.012496 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.012871 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.013653 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.016326 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.017738 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.018375 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.019223 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.020564 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.021151 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.022364 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.023606 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.024340 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.025327 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.026074 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.027116 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.028222 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.029335 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.029693 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.029978 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.030547 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.031688 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.032394 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.033422 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.048263 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.106983 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"853e1a511b94837123c4e1f9a8bc46b089a55efc7b1447e96ed9051fcefa14bf"} Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.108030 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80"} Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.108062 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ca49f4345a8dd9a2289278beb257ee8a8ccc3a68bf877395666637f0ebd6a520"} Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.109624 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398"} Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.109648 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0"} Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.109659 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d46fa10305ae36d7eb191372b80e0f3c2dbed0a5abf8da2fca12df53f7e1b262"} Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.111639 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.113061 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b"} Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.113496 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.122452 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.136482 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.156765 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.181496 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.199728 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.224742 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.241107 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.265561 4677 csr.go:261] certificate signing request csr-8hzjz is approved, waiting to be issued Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.269799 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.278691 4677 csr.go:257] certificate signing request csr-8hzjz is issued Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.316477 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.369548 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.517645 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.517782 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.517868 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.517870 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:47:12.517833576 +0000 UTC m=+23.264166191 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.518017 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:12.518004072 +0000 UTC m=+23.264336747 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.619033 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.619070 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.619103 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619235 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619285 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619299 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619351 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:12.619335315 +0000 UTC m=+23.365667770 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619344 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619441 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:12.619419338 +0000 UTC m=+23.365751853 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619241 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619478 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619493 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.619525 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:12.619517311 +0000 UTC m=+23.365849866 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.679738 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-xjvnv"] Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.680060 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.682001 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.682409 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.682584 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.686064 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-2bx7x"] Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.686430 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.688475 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.688895 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.690178 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.690825 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.691571 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.698442 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.714995 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.719631 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1665842-4668-4fed-a04f-cf4aa0043ebc-proxy-tls\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.719668 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kv6j\" (UniqueName: \"kubernetes.io/projected/76e85245-8632-41e9-b4bd-987b708709a8-kube-api-access-9kv6j\") pod \"node-resolver-xjvnv\" (UID: \"76e85245-8632-41e9-b4bd-987b708709a8\") " pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.719695 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1665842-4668-4fed-a04f-cf4aa0043ebc-mcd-auth-proxy-config\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.719743 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/76e85245-8632-41e9-b4bd-987b708709a8-hosts-file\") pod \"node-resolver-xjvnv\" (UID: \"76e85245-8632-41e9-b4bd-987b708709a8\") " pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.719765 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c1665842-4668-4fed-a04f-cf4aa0043ebc-rootfs\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.719816 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf2m2\" (UniqueName: \"kubernetes.io/projected/c1665842-4668-4fed-a04f-cf4aa0043ebc-kube-api-access-qf2m2\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.729043 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.744275 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.756887 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.773687 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821115 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821487 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kv6j\" (UniqueName: \"kubernetes.io/projected/76e85245-8632-41e9-b4bd-987b708709a8-kube-api-access-9kv6j\") pod \"node-resolver-xjvnv\" (UID: \"76e85245-8632-41e9-b4bd-987b708709a8\") " pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821545 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1665842-4668-4fed-a04f-cf4aa0043ebc-mcd-auth-proxy-config\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821593 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/76e85245-8632-41e9-b4bd-987b708709a8-hosts-file\") pod \"node-resolver-xjvnv\" (UID: \"76e85245-8632-41e9-b4bd-987b708709a8\") " pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821619 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c1665842-4668-4fed-a04f-cf4aa0043ebc-rootfs\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821653 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf2m2\" (UniqueName: \"kubernetes.io/projected/c1665842-4668-4fed-a04f-cf4aa0043ebc-kube-api-access-qf2m2\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821681 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1665842-4668-4fed-a04f-cf4aa0043ebc-proxy-tls\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.821800 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c1665842-4668-4fed-a04f-cf4aa0043ebc-rootfs\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.822152 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/76e85245-8632-41e9-b4bd-987b708709a8-hosts-file\") pod \"node-resolver-xjvnv\" (UID: \"76e85245-8632-41e9-b4bd-987b708709a8\") " pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.823010 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1665842-4668-4fed-a04f-cf4aa0043ebc-mcd-auth-proxy-config\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.833422 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1665842-4668-4fed-a04f-cf4aa0043ebc-proxy-tls\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.839836 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.847082 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf2m2\" (UniqueName: \"kubernetes.io/projected/c1665842-4668-4fed-a04f-cf4aa0043ebc-kube-api-access-qf2m2\") pod \"machine-config-daemon-2bx7x\" (UID: \"c1665842-4668-4fed-a04f-cf4aa0043ebc\") " pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.855495 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kv6j\" (UniqueName: \"kubernetes.io/projected/76e85245-8632-41e9-b4bd-987b708709a8-kube-api-access-9kv6j\") pod \"node-resolver-xjvnv\" (UID: \"76e85245-8632-41e9-b4bd-987b708709a8\") " pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.874223 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.928089 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.954660 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.975501 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.975571 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.975623 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:10 crc kubenswrapper[4677]: E1203 13:47:10.975692 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.986908 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:10 crc kubenswrapper[4677]: I1203 13:47:10.993069 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xjvnv" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.002268 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:47:11 crc kubenswrapper[4677]: W1203 13:47:11.006914 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76e85245_8632_41e9_b4bd_987b708709a8.slice/crio-f6f0d369119dbaabb20f78ecdd55c6e00d28bfd5caf92c00dd65035303e47c18 WatchSource:0}: Error finding container f6f0d369119dbaabb20f78ecdd55c6e00d28bfd5caf92c00dd65035303e47c18: Status 404 returned error can't find the container with id f6f0d369119dbaabb20f78ecdd55c6e00d28bfd5caf92c00dd65035303e47c18 Dec 03 13:47:11 crc kubenswrapper[4677]: W1203 13:47:11.014897 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1665842_4668_4fed_a04f_cf4aa0043ebc.slice/crio-20791284a46563e2ce5a367683f700229cce235a101335e48f296e3eb39fd4c6 WatchSource:0}: Error finding container 20791284a46563e2ce5a367683f700229cce235a101335e48f296e3eb39fd4c6: Status 404 returned error can't find the container with id 20791284a46563e2ce5a367683f700229cce235a101335e48f296e3eb39fd4c6 Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.027916 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.094315 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-7nch7"] Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.094556 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-sxc65"] Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.094739 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.095329 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pm7pz"] Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.095528 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.096686 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.101647 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.101709 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.101742 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.101888 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.102044 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.102282 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.102413 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.102517 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.105229 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.105278 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.105342 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.105648 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.145201 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.145677 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.201055 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.203924 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"20791284a46563e2ce5a367683f700229cce235a101335e48f296e3eb39fd4c6"} Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.210767 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xjvnv" event={"ID":"76e85245-8632-41e9-b4bd-987b708709a8","Type":"ContainerStarted","Data":"f6f0d369119dbaabb20f78ecdd55c6e00d28bfd5caf92c00dd65035303e47c18"} Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.224908 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.237540 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.244306 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-cni-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.244521 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaf2e40d-0316-4380-961f-8039e0674f2c-ovn-node-metrics-cert\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.244663 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.244750 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-etc-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.244850 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-cni-bin\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.244923 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-hostroot\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245012 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-netns\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245075 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-os-release\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245135 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cni-binary-copy\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245246 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245295 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-ovn\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245317 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-bin\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245336 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-config\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245393 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-os-release\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245412 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/511a6f65-7dac-4f37-a15e-3a24339f80f4-cni-binary-copy\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245432 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-system-cni-dir\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245456 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhct7\" (UniqueName: \"kubernetes.io/projected/511a6f65-7dac-4f37-a15e-3a24339f80f4-kube-api-access-qhct7\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245481 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-systemd\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245513 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-socket-dir-parent\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245533 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-log-socket\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245556 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lxth\" (UniqueName: \"kubernetes.io/projected/eaf2e40d-0316-4380-961f-8039e0674f2c-kube-api-access-9lxth\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245578 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-k8s-cni-cncf-io\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245596 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-daemon-config\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245615 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-script-lib\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245636 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-cnibin\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.245656 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-conf-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.247842 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-node-log\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248174 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248256 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-env-overrides\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248288 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-systemd-units\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248354 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-slash\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248402 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-var-lib-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248428 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-cni-multus\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248510 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-multus-certs\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248535 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-kubelet\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248555 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248631 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-netns\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248653 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-etc-kubernetes\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248679 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248705 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-kubelet\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248782 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-netd\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248864 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-system-cni-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248897 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cnibin\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.248927 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnvgv\" (UniqueName: \"kubernetes.io/projected/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-kube-api-access-fnvgv\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.255700 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.280187 4677 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-03 13:42:10 +0000 UTC, rotation deadline is 2026-09-07 19:16:48.891802681 +0000 UTC Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.280265 4677 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6677h29m37.611542226s for next certificate rotation Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.288129 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.307789 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.328108 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.343864 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.350974 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-cni-multus\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351016 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-multus-certs\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351038 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-kubelet\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351059 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-netns\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351076 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-etc-kubernetes\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351073 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-cni-multus\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351165 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351094 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351198 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-netns\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351215 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351290 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-kubelet\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351310 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-netd\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351328 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-system-cni-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351348 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cnibin\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351363 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnvgv\" (UniqueName: \"kubernetes.io/projected/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-kube-api-access-fnvgv\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351381 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-cni-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351396 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaf2e40d-0316-4380-961f-8039e0674f2c-ovn-node-metrics-cert\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351410 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351439 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-cni-bin\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351455 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-hostroot\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351470 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-netns\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351496 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-etc-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351526 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351541 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-ovn\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351559 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-os-release\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351600 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cni-binary-copy\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351620 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-bin\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351635 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-config\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351676 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-os-release\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351706 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/511a6f65-7dac-4f37-a15e-3a24339f80f4-cni-binary-copy\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351737 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-system-cni-dir\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351755 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhct7\" (UniqueName: \"kubernetes.io/projected/511a6f65-7dac-4f37-a15e-3a24339f80f4-kube-api-access-qhct7\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351771 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-systemd\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351797 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-socket-dir-parent\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351812 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-log-socket\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351828 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lxth\" (UniqueName: \"kubernetes.io/projected/eaf2e40d-0316-4380-961f-8039e0674f2c-kube-api-access-9lxth\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351843 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-k8s-cni-cncf-io\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351858 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-daemon-config\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351872 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-script-lib\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351882 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351888 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-cnibin\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351925 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-cnibin\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351926 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-conf-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351964 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-conf-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351970 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-node-log\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351989 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-systemd-units\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351992 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-ovn\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351215 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-etc-kubernetes\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352006 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-slash\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352029 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-var-lib-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352047 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352065 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-env-overrides\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352272 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-os-release\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352527 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-env-overrides\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352572 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-kubelet\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.351192 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-multus-certs\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352618 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-netd\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352739 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-system-cni-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352774 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cnibin\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352846 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-cni-binary-copy\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352889 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-bin\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353046 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-log-socket\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353135 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353135 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-etc-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353180 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-var-lib-openvswitch\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.352030 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-slash\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353209 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353341 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-os-release\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353402 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-config\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353355 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-node-log\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353463 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-systemd-units\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353519 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-systemd\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353574 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-system-cni-dir\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353853 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-daemon-config\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353883 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-script-lib\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353927 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-run-k8s-cni-cncf-io\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353961 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-kubelet\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353975 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-host-var-lib-cni-bin\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.354006 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-hostroot\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.353069 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-netns\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.354017 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-socket-dir-parent\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.354450 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/511a6f65-7dac-4f37-a15e-3a24339f80f4-cni-binary-copy\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.355025 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/511a6f65-7dac-4f37-a15e-3a24339f80f4-multus-cni-dir\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.356830 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.357631 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaf2e40d-0316-4380-961f-8039e0674f2c-ovn-node-metrics-cert\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.357994 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.375421 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.375841 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhct7\" (UniqueName: \"kubernetes.io/projected/511a6f65-7dac-4f37-a15e-3a24339f80f4-kube-api-access-qhct7\") pod \"multus-7nch7\" (UID: \"511a6f65-7dac-4f37-a15e-3a24339f80f4\") " pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.376605 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnvgv\" (UniqueName: \"kubernetes.io/projected/b0755a37-6619-4dd8-b598-d2bf05c3a3ab-kube-api-access-fnvgv\") pod \"multus-additional-cni-plugins-sxc65\" (UID: \"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\") " pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.377890 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lxth\" (UniqueName: \"kubernetes.io/projected/eaf2e40d-0316-4380-961f-8039e0674f2c-kube-api-access-9lxth\") pod \"ovnkube-node-pm7pz\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.396512 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.409835 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7nch7" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.409891 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.441374 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.447039 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sxc65" Dec 03 13:47:11 crc kubenswrapper[4677]: W1203 13:47:11.465005 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0755a37_6619_4dd8_b598_d2bf05c3a3ab.slice/crio-c31ac533fd1c1cb4dc9854ec453f2ac00de64d04b50d1a07f569efaec94c0b98 WatchSource:0}: Error finding container c31ac533fd1c1cb4dc9854ec453f2ac00de64d04b50d1a07f569efaec94c0b98: Status 404 returned error can't find the container with id c31ac533fd1c1cb4dc9854ec453f2ac00de64d04b50d1a07f569efaec94c0b98 Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.469115 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.486300 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.505156 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:11Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.510901 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:11 crc kubenswrapper[4677]: W1203 13:47:11.532136 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaf2e40d_0316_4380_961f_8039e0674f2c.slice/crio-5a56617fdb324443c399a5eaf6621a4a2d5db7865adccb41b08455ba73d584c3 WatchSource:0}: Error finding container 5a56617fdb324443c399a5eaf6621a4a2d5db7865adccb41b08455ba73d584c3: Status 404 returned error can't find the container with id 5a56617fdb324443c399a5eaf6621a4a2d5db7865adccb41b08455ba73d584c3 Dec 03 13:47:11 crc kubenswrapper[4677]: I1203 13:47:11.975429 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:11 crc kubenswrapper[4677]: E1203 13:47:11.975914 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.226784 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerStarted","Data":"1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.226882 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerStarted","Data":"c31ac533fd1c1cb4dc9854ec453f2ac00de64d04b50d1a07f569efaec94c0b98"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.228522 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde" exitCode=0 Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.228624 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.228663 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"5a56617fdb324443c399a5eaf6621a4a2d5db7865adccb41b08455ba73d584c3"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.230250 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.251265 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.254675 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerStarted","Data":"0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.254719 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerStarted","Data":"928a0354286e5d7b17b7e20139626433485945c3f103eefb0d40d105fb594026"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.257874 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.257926 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.259137 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xjvnv" event={"ID":"76e85245-8632-41e9-b4bd-987b708709a8","Type":"ContainerStarted","Data":"ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.282767 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.288420 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.299240 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.301874 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.302081 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.313044 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.334143 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.348159 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.367613 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.387321 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.403742 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.423627 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.441214 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.457357 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.475090 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.493263 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.515341 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.540056 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.553370 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.569280 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.569331 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.569428 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:47:16.56940329 +0000 UTC m=+27.315735745 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.569484 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.569597 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.569648 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:16.569636287 +0000 UTC m=+27.315968742 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.583077 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.595119 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.606564 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.621881 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.632548 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.644728 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.657806 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.670408 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.670462 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.670504 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670618 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670633 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670668 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670684 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670695 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:16.670676302 +0000 UTC m=+27.417008757 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670622 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670760 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670776 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670735 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:16.670718083 +0000 UTC m=+27.417050618 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.670829 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:16.670815276 +0000 UTC m=+27.417147831 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.763625 4677 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.765299 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.765331 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.765343 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.765407 4677 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.771874 4677 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.772110 4677 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.773294 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.773321 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.773332 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.773348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.773359 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:12Z","lastTransitionTime":"2025-12-03T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.792568 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.795915 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.795963 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.795977 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.795994 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.796005 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:12Z","lastTransitionTime":"2025-12-03T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.808125 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.812078 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.812119 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.812138 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.812237 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.812250 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:12Z","lastTransitionTime":"2025-12-03T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.825826 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.829202 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.829241 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.829254 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.829272 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.829284 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:12Z","lastTransitionTime":"2025-12-03T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.842586 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.845811 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.845849 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.845861 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.845879 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.845891 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:12Z","lastTransitionTime":"2025-12-03T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.859428 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:12Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.859562 4677 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.861483 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.861540 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.861552 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.861571 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.861591 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:12Z","lastTransitionTime":"2025-12-03T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.964596 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.964648 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.964658 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.964673 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.964682 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:12Z","lastTransitionTime":"2025-12-03T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.975041 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:12 crc kubenswrapper[4677]: I1203 13:47:12.975058 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.975165 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:12 crc kubenswrapper[4677]: E1203 13:47:12.975219 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.067238 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.067267 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.067276 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.067290 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.067300 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.170324 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.170717 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.170729 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.170750 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.170760 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.266886 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.266933 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.266964 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.266975 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.269266 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0755a37-6619-4dd8-b598-d2bf05c3a3ab" containerID="1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977" exitCode=0 Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.269742 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerDied","Data":"1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.274330 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.274376 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.274389 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.274407 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.274418 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: E1203 13:47:13.280883 4677 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.288003 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.308420 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.326526 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.346022 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.359355 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.370124 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.377856 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.377935 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.378003 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.378035 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.378188 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.382264 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.397338 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.409456 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.428256 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.441439 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.454820 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.467695 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.480757 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.480808 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.480820 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.480838 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.480849 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.529366 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.533046 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.537123 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.540654 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.553215 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.568883 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.581384 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.582912 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.582993 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.583003 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.583019 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.583028 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.594207 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.605114 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.625218 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.670086 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.685320 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.685357 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.685366 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.685380 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.685389 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.688313 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.708925 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.723704 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.736408 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.756538 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.770546 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.781343 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.787290 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.787323 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.787330 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.787344 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.787352 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.795087 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.813356 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.813466 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-bdbzd"] Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.813823 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.815206 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.816446 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.816564 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.816671 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.828408 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.841132 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.858889 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.873576 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.899408 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.899620 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2cfbbe23-6cec-425e-a192-10c5f7f36317-host\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.899665 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2cfbbe23-6cec-425e-a192-10c5f7f36317-serviceca\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.899697 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn7qj\" (UniqueName: \"kubernetes.io/projected/2cfbbe23-6cec-425e-a192-10c5f7f36317-kube-api-access-zn7qj\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.901113 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.901146 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.901157 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.901174 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.901185 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:13Z","lastTransitionTime":"2025-12-03T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.952971 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.970448 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.976065 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:13 crc kubenswrapper[4677]: E1203 13:47:13.976229 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.986049 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:13 crc kubenswrapper[4677]: I1203 13:47:13.999238 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:13Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.000494 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2cfbbe23-6cec-425e-a192-10c5f7f36317-host\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.000522 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2cfbbe23-6cec-425e-a192-10c5f7f36317-serviceca\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.000564 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn7qj\" (UniqueName: \"kubernetes.io/projected/2cfbbe23-6cec-425e-a192-10c5f7f36317-kube-api-access-zn7qj\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.000605 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2cfbbe23-6cec-425e-a192-10c5f7f36317-host\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.001496 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2cfbbe23-6cec-425e-a192-10c5f7f36317-serviceca\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.002775 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.002809 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.002821 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.002836 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.002846 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.018120 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn7qj\" (UniqueName: \"kubernetes.io/projected/2cfbbe23-6cec-425e-a192-10c5f7f36317-kube-api-access-zn7qj\") pod \"node-ca-bdbzd\" (UID: \"2cfbbe23-6cec-425e-a192-10c5f7f36317\") " pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.023823 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.036558 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.046295 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.059233 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.072158 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.085093 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.105665 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.105710 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.105727 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.105744 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.105755 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.109394 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.118902 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.125167 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bdbzd" Dec 03 13:47:14 crc kubenswrapper[4677]: W1203 13:47:14.138370 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cfbbe23_6cec_425e_a192_10c5f7f36317.slice/crio-ab66bc7e231b057b461ca8bc4d92e541fe3a11a46bbcbe4a78bb574433ba0e7f WatchSource:0}: Error finding container ab66bc7e231b057b461ca8bc4d92e541fe3a11a46bbcbe4a78bb574433ba0e7f: Status 404 returned error can't find the container with id ab66bc7e231b057b461ca8bc4d92e541fe3a11a46bbcbe4a78bb574433ba0e7f Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.139751 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.155001 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.172091 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.188440 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.217221 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.217285 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.217298 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.217320 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.217340 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.218552 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.234554 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.250312 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.273168 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.278314 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerStarted","Data":"c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.286088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.286474 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.290292 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bdbzd" event={"ID":"2cfbbe23-6cec-425e-a192-10c5f7f36317","Type":"ContainerStarted","Data":"ab66bc7e231b057b461ca8bc4d92e541fe3a11a46bbcbe4a78bb574433ba0e7f"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.320126 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.320190 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.320203 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.320226 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.320242 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.335335 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.350671 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.362321 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.382116 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.396201 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.408551 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.421919 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.436203 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.440090 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.440142 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.440153 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.440170 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.440184 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.449453 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.465603 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.481914 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.497847 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.515005 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.528740 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.539700 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:14Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.542708 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.542745 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.542754 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.542771 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.542784 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.645819 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.645856 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.645865 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.645879 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.645887 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.748095 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.748134 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.748142 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.748158 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.748170 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.851326 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.851571 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.851583 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.851598 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.851608 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.954288 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.954324 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.954334 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.954348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.954357 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:14Z","lastTransitionTime":"2025-12-03T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.975762 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:14 crc kubenswrapper[4677]: I1203 13:47:14.975804 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:14 crc kubenswrapper[4677]: E1203 13:47:14.975880 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:14 crc kubenswrapper[4677]: E1203 13:47:14.975967 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.056259 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.056300 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.056309 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.056323 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.056333 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.158528 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.158569 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.158582 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.158601 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.158613 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.260653 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.260703 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.260713 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.260728 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.260737 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.294680 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bdbzd" event={"ID":"2cfbbe23-6cec-425e-a192-10c5f7f36317","Type":"ContainerStarted","Data":"2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.296566 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0755a37-6619-4dd8-b598-d2bf05c3a3ab" containerID="c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c" exitCode=0 Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.296617 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerDied","Data":"c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.308436 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.322531 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.335489 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.358125 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.365314 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.365382 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.365396 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.365415 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.365426 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.371622 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.391775 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.414002 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.437110 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.456671 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.467568 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.467603 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.467614 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.467630 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.467642 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.473128 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.484166 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.500918 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.512664 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.531174 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.548066 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.572247 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.573605 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.573633 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.573643 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.573661 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.573673 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.588887 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.601723 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.618528 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.626879 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.637616 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.647694 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.660027 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.673623 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.675454 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.675484 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.675497 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.675515 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.675526 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.685073 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.694033 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.704851 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.715845 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.727045 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.737094 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:15Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.777164 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.777194 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.777228 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.777241 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.777250 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.880131 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.880184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.880199 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.880222 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.880250 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.975055 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:15 crc kubenswrapper[4677]: E1203 13:47:15.975189 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.982621 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.982664 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.982674 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.982689 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:15 crc kubenswrapper[4677]: I1203 13:47:15.982701 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:15Z","lastTransitionTime":"2025-12-03T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.084691 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.084731 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.084743 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.084758 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.084768 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.187247 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.187274 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.187285 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.187299 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.187310 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.289213 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.289253 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.289262 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.289276 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.289286 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.301440 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0755a37-6619-4dd8-b598-d2bf05c3a3ab" containerID="3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7" exitCode=0 Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.301506 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerDied","Data":"3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.307925 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.315374 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.334549 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.351236 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.364326 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.377691 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.392417 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.392457 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.392468 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.392486 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.392500 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.392703 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.406981 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.422698 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.444162 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.467017 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.481125 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.491983 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.494514 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.494552 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.494566 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.494583 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.494595 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.508360 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.523385 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.536452 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:16Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.597235 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.597283 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.597299 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.597321 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.597338 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.624759 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.624864 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.625029 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.625085 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:47:24.625050649 +0000 UTC m=+35.371383124 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.625145 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:24.625117671 +0000 UTC m=+35.371450136 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.700113 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.700155 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.700184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.700204 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.700216 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.725658 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.725714 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.725754 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725840 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725850 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725872 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725885 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725907 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:24.725889898 +0000 UTC m=+35.472222353 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725924 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:24.725915199 +0000 UTC m=+35.472247654 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725943 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.725988 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.726001 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.726063 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:24.726046873 +0000 UTC m=+35.472379328 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.802871 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.802919 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.802934 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.802992 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.803009 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.906017 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.906069 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.906082 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.906102 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.906115 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:16Z","lastTransitionTime":"2025-12-03T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.976057 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.976214 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:16 crc kubenswrapper[4677]: I1203 13:47:16.976504 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:16 crc kubenswrapper[4677]: E1203 13:47:16.976721 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.009463 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.009521 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.009533 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.009552 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.009567 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.112388 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.112449 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.112462 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.112482 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.112495 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.215187 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.215434 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.215442 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.215457 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.215467 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.317363 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0755a37-6619-4dd8-b598-d2bf05c3a3ab" containerID="83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00" exitCode=0 Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.317415 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerDied","Data":"83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.319456 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.319479 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.319487 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.319501 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.319511 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.333617 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.349216 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.368288 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.380338 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.390933 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.409290 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.421684 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.421730 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.421740 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.421756 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.421772 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.431991 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.444610 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.457939 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.471388 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.481976 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.493251 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.505263 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.516041 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.524620 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.524658 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.524670 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.524688 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.524702 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.526039 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:17Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.630059 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.630094 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.630102 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.630118 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.630126 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.732967 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.733014 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.733025 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.733040 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.733052 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.835906 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.835972 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.835985 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.836003 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.836017 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.938730 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.938769 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.938778 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.938794 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.938803 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:17Z","lastTransitionTime":"2025-12-03T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:17 crc kubenswrapper[4677]: I1203 13:47:17.982831 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:17 crc kubenswrapper[4677]: E1203 13:47:17.982976 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.041395 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.041428 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.041439 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.041453 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.041463 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.144887 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.144934 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.144960 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.144976 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.144984 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.247517 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.247569 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.247578 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.247594 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.247602 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.323059 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerStarted","Data":"5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.327156 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.328327 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.328383 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.344385 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.350052 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.350092 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.350103 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.350123 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.350134 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.355910 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.367021 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.384080 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.394526 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.420147 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.431249 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.437296 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.438979 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.450233 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.452012 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.452047 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.452058 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.452073 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.452087 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.463842 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.474693 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.489279 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.505294 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.518202 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.532851 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.548086 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.556281 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.556684 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.556764 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.556831 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.556894 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.564474 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.578068 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.596259 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.611116 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.627527 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.642232 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.657542 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.660318 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.660348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.660360 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.660383 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.660396 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.671674 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.691567 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.712466 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.727233 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.744910 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.761059 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.763719 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.763757 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.763769 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.763793 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.763806 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.776938 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.790528 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:18Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.867424 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.867980 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.867996 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.868021 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.868036 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.972292 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.972349 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.972361 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.972384 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.972398 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:18Z","lastTransitionTime":"2025-12-03T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.975748 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:18 crc kubenswrapper[4677]: I1203 13:47:18.975822 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:18 crc kubenswrapper[4677]: E1203 13:47:18.975941 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:18 crc kubenswrapper[4677]: E1203 13:47:18.976071 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.075867 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.075912 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.075925 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.075961 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.075994 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.178765 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.178805 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.178814 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.178827 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.178837 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.281332 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.282015 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.282033 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.282063 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.282076 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.335097 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0755a37-6619-4dd8-b598-d2bf05c3a3ab" containerID="5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3" exitCode=0 Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.335154 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerDied","Data":"5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.335304 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.353066 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.369090 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.386825 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.386875 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.386885 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.386902 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.386914 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.387225 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.405369 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.425158 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.440928 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.463017 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.484499 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.489490 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.489534 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.489543 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.489558 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.489567 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.501576 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.516047 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.533517 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.546577 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.563344 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.576718 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.591583 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.592146 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.592181 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.592192 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.592207 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.592218 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.697483 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.697526 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.697555 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.697572 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.697583 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.764789 4677 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 03 13:47:19 crc kubenswrapper[4677]: E1203 13:47:19.765230 4677 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/events\": read tcp 38.102.83.162:45266->38.102.83.162:6443: use of closed network connection" event="&Event{ObjectMeta:{multus-additional-cni-plugins-sxc65.187db8a3206b1877 openshift-multus 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-multus,Name:multus-additional-cni-plugins-sxc65,UID:b0755a37-6619-4dd8-b598-d2bf05c3a3ab,APIVersion:v1,ResourceVersion:26625,FieldPath:spec.initContainers{whereabouts-cni},},Reason:Created,Message:Created container whereabouts-cni,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:47:19.762172023 +0000 UTC m=+30.508504488,LastTimestamp:2025-12-03 13:47:19.762172023 +0000 UTC m=+30.508504488,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.800660 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.800807 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.800873 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.800941 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.801052 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.904195 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.904240 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.904251 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.904271 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.904286 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:19Z","lastTransitionTime":"2025-12-03T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.975586 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:19 crc kubenswrapper[4677]: E1203 13:47:19.975725 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:19 crc kubenswrapper[4677]: I1203 13:47:19.990223 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:19Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.004407 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.007719 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.008066 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.008350 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.008439 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.008527 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.029235 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.049015 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.062131 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.082854 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.097817 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.142094 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.142142 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.142151 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.142166 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.142177 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.153283 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.174219 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.190356 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.208639 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.226851 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.245174 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.245446 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.245557 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.245704 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.245793 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.251389 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.273152 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.291371 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.348168 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.348212 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.348224 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.348242 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.348254 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.349829 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0755a37-6619-4dd8-b598-d2bf05c3a3ab" containerID="61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532" exitCode=0 Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.350060 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerDied","Data":"61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.350622 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.369164 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.390279 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.410005 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.424063 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.444862 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.451332 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.451360 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.451370 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.451382 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.451393 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.457771 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.485179 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.502311 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.519579 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.539165 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.554486 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.554519 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.554529 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.554544 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.554555 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.556075 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.569853 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.583835 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.601925 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.613827 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:20Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.657096 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.657143 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.657153 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.657169 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.657179 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.760390 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.760827 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.760840 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.760859 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.760875 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.863778 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.863844 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.863858 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.863881 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.863897 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.966715 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.966747 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.966755 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.966769 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.966778 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:20Z","lastTransitionTime":"2025-12-03T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.975147 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:20 crc kubenswrapper[4677]: E1203 13:47:20.975313 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:20 crc kubenswrapper[4677]: I1203 13:47:20.975465 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:20 crc kubenswrapper[4677]: E1203 13:47:20.975689 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.069085 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.069116 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.069125 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.069138 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.069147 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.172310 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.172584 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.172708 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.172842 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.172974 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.275413 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.275668 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.275781 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.275866 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.275968 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.356161 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" event={"ID":"b0755a37-6619-4dd8-b598-d2bf05c3a3ab","Type":"ContainerStarted","Data":"ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.358120 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/0.log" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.361133 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3" exitCode=1 Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.361385 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.366657 4677 scope.go:117] "RemoveContainer" containerID="7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.375009 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.378093 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.378133 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.378146 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.378167 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.378180 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.391735 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.408110 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.422386 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.435876 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.453394 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.473615 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.480752 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.480783 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.480793 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.480808 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.480817 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.496070 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.510466 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.525931 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.543117 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.558272 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.577830 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.582530 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.582599 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.582612 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.582629 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.582641 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.592712 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.607780 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.621824 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.654196 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.684680 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.684714 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.684723 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.684736 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.684747 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.690819 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.709050 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.721172 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.733983 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.744926 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.759127 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.772029 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.787427 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.787460 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.787467 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.787480 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.787489 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.788963 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.805728 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.820639 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.835539 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.852931 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.876007 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:21Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.889465 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.889501 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.889511 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.889525 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.889536 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.975398 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:21 crc kubenswrapper[4677]: E1203 13:47:21.975520 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.991901 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.991943 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.991969 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.991987 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:21 crc kubenswrapper[4677]: I1203 13:47:21.992000 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:21Z","lastTransitionTime":"2025-12-03T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.094444 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.094477 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.094485 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.094501 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.094510 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.196634 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.196676 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.196687 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.196733 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.196745 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.299378 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.299428 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.299441 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.299460 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.299472 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.366918 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/0.log" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.369071 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.369193 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.384714 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.397517 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.402128 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.402166 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.402180 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.402200 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.402214 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.421319 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.449405 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.462842 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.476768 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.493193 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.504638 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.504761 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.504807 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.504824 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.504846 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.504864 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.518901 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.534639 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.544528 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.556769 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.568233 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.581553 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.592785 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.610404 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.610900 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.610923 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.610941 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.610980 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.714888 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.715016 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.715038 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.715070 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.715089 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.818189 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.818245 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.818264 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.818285 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.818303 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.890605 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.890759 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.890788 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.890821 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.890845 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: E1203 13:47:22.915008 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.922141 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.922195 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.922214 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.922238 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.922256 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: E1203 13:47:22.941015 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.946678 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.946759 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.946783 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.946811 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.946829 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: E1203 13:47:22.964743 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.970320 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.970382 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.970402 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.970427 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.970449 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.975387 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.975398 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:22 crc kubenswrapper[4677]: E1203 13:47:22.975554 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:22 crc kubenswrapper[4677]: E1203 13:47:22.975725 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:22 crc kubenswrapper[4677]: E1203 13:47:22.987615 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:22Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.992367 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.992429 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.992450 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.992475 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:22 crc kubenswrapper[4677]: I1203 13:47:22.992495 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:22Z","lastTransitionTime":"2025-12-03T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: E1203 13:47:23.005534 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: E1203 13:47:23.005839 4677 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.008102 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.008160 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.008185 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.008210 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.008233 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.111523 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.111593 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.111616 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.111645 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.111668 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.213507 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.213553 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.213564 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.213587 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.213598 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.315937 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.315997 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.316022 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.316044 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.316058 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.375771 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/1.log" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.376937 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/0.log" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.381098 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b" exitCode=1 Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.381153 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.381239 4677 scope.go:117] "RemoveContainer" containerID="7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.382498 4677 scope.go:117] "RemoveContainer" containerID="008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b" Dec 03 13:47:23 crc kubenswrapper[4677]: E1203 13:47:23.382815 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.418827 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.419504 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.419580 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.419605 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.419635 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.419662 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.442081 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.462645 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.492679 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.515716 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.522386 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.522468 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.522489 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.522574 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.522594 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.533519 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.549125 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.566122 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.620678 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.625882 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.625930 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.625956 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.625979 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.625991 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.637111 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.650097 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.669330 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.684698 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.705535 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.723836 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.728811 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.728852 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.728863 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.728882 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.728896 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.832105 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.832168 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.832190 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.832216 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.832235 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.912407 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql"] Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.913664 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.917057 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.917110 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.935937 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.936048 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.936068 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.936095 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.936116 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:23Z","lastTransitionTime":"2025-12-03T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.946247 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.964122 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.975301 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:23 crc kubenswrapper[4677]: E1203 13:47:23.975466 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:23 crc kubenswrapper[4677]: I1203 13:47:23.980286 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:23.999805 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:23Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.012718 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj2m5\" (UniqueName: \"kubernetes.io/projected/97274f10-85e1-4b88-b061-36427e54a9d4-kube-api-access-nj2m5\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.012782 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97274f10-85e1-4b88-b061-36427e54a9d4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.012836 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97274f10-85e1-4b88-b061-36427e54a9d4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.012880 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97274f10-85e1-4b88-b061-36427e54a9d4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.013824 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.030124 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.039167 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.039203 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.039214 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.039230 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.039242 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.056600 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.071733 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.084454 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.106503 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.113444 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj2m5\" (UniqueName: \"kubernetes.io/projected/97274f10-85e1-4b88-b061-36427e54a9d4-kube-api-access-nj2m5\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.113510 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97274f10-85e1-4b88-b061-36427e54a9d4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.113583 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97274f10-85e1-4b88-b061-36427e54a9d4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.113627 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97274f10-85e1-4b88-b061-36427e54a9d4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.114372 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97274f10-85e1-4b88-b061-36427e54a9d4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.114678 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97274f10-85e1-4b88-b061-36427e54a9d4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.118855 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.120763 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97274f10-85e1-4b88-b061-36427e54a9d4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.132675 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.133825 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj2m5\" (UniqueName: \"kubernetes.io/projected/97274f10-85e1-4b88-b061-36427e54a9d4-kube-api-access-nj2m5\") pod \"ovnkube-control-plane-749d76644c-mjkql\" (UID: \"97274f10-85e1-4b88-b061-36427e54a9d4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.141324 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.141381 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.141399 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.141420 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.141431 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.148675 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.161768 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.174785 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.187489 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:24Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.232257 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.243724 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.243765 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.243781 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.243804 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.243820 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: W1203 13:47:24.250505 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97274f10_85e1_4b88_b061_36427e54a9d4.slice/crio-2d31592cf0a23e2d40c3250f5c03f079bed9f2b8f306d70c02d329c2ee25a042 WatchSource:0}: Error finding container 2d31592cf0a23e2d40c3250f5c03f079bed9f2b8f306d70c02d329c2ee25a042: Status 404 returned error can't find the container with id 2d31592cf0a23e2d40c3250f5c03f079bed9f2b8f306d70c02d329c2ee25a042 Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.346126 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.346165 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.346175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.346189 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.346197 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.385807 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" event={"ID":"97274f10-85e1-4b88-b061-36427e54a9d4","Type":"ContainerStarted","Data":"2d31592cf0a23e2d40c3250f5c03f079bed9f2b8f306d70c02d329c2ee25a042"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.387467 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/1.log" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.448773 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.448819 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.448832 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.448851 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.448863 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.553028 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.553081 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.553097 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.553120 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.553145 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.656882 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.656962 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.656971 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.656986 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.656994 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.720592 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.720751 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:47:40.720727036 +0000 UTC m=+51.467059491 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.720887 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.720993 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.721039 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:40.721032936 +0000 UTC m=+51.467365391 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.759521 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.759551 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.759559 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.759573 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.759581 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.821679 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.821738 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.821768 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.821889 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.821906 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.821916 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.821982 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:40.821965808 +0000 UTC m=+51.568298263 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.822046 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.822057 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.822065 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.822092 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:40.822084241 +0000 UTC m=+51.568416696 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.822191 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.822277 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:40.822259496 +0000 UTC m=+51.568591951 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.862575 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.862633 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.862647 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.862668 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.862681 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.965513 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.965559 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.965571 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.965589 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.965600 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:24Z","lastTransitionTime":"2025-12-03T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.975771 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:24 crc kubenswrapper[4677]: I1203 13:47:24.975791 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.975903 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:24 crc kubenswrapper[4677]: E1203 13:47:24.976001 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.018203 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-j7h62"] Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.018596 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:25 crc kubenswrapper[4677]: E1203 13:47:25.018652 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.029729 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.041715 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.056689 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.066905 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.067745 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.067786 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.067799 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.067816 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.067827 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.080820 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.096478 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.110299 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.121176 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.124898 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcgcc\" (UniqueName: \"kubernetes.io/projected/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-kube-api-access-bcgcc\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.125103 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.143553 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.156809 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.168716 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.170448 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.170496 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.170511 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.170528 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.170541 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.187876 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.201275 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.214527 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.225783 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.225842 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcgcc\" (UniqueName: \"kubernetes.io/projected/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-kube-api-access-bcgcc\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:25 crc kubenswrapper[4677]: E1203 13:47:25.225973 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:25 crc kubenswrapper[4677]: E1203 13:47:25.226026 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:25.726011953 +0000 UTC m=+36.472344408 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.226609 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.241547 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcgcc\" (UniqueName: \"kubernetes.io/projected/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-kube-api-access-bcgcc\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.245798 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.270616 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.272495 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.272531 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.272541 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.272555 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.272563 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.375469 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.375520 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.375532 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.375549 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.375560 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.400528 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" event={"ID":"97274f10-85e1-4b88-b061-36427e54a9d4","Type":"ContainerStarted","Data":"544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.400584 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" event={"ID":"97274f10-85e1-4b88-b061-36427e54a9d4","Type":"ContainerStarted","Data":"51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.416567 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.432138 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.447219 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.460549 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.471456 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.478180 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.478227 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.478246 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.478266 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.478283 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.491335 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.505441 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.519097 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.535631 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.547483 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.561682 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.574697 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.580648 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.580706 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.580719 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.580736 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.580755 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.589119 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.603198 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.618489 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.634667 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.650171 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:25Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.687167 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.687204 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.687212 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.687225 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.687236 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.730859 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:25 crc kubenswrapper[4677]: E1203 13:47:25.731124 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:25 crc kubenswrapper[4677]: E1203 13:47:25.731419 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:26.731398473 +0000 UTC m=+37.477730918 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.789694 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.789724 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.789735 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.789752 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.789761 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.892873 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.892907 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.892917 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.892929 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.892939 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.975161 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:25 crc kubenswrapper[4677]: E1203 13:47:25.975349 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.995571 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.995637 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.995667 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.995697 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:25 crc kubenswrapper[4677]: I1203 13:47:25.995721 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:25Z","lastTransitionTime":"2025-12-03T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.098931 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.098987 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.098999 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.099017 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.099029 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.202346 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.202385 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.202396 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.202410 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.202427 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.305393 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.305666 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.305743 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.305831 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.305902 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.409101 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.409193 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.409268 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.409294 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.409312 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.511395 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.511462 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.511494 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.511512 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.511523 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.615663 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.615733 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.615747 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.615772 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.615787 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.718255 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.718320 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.718336 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.718360 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.718376 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.740669 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:26 crc kubenswrapper[4677]: E1203 13:47:26.740834 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:26 crc kubenswrapper[4677]: E1203 13:47:26.740923 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:28.740899816 +0000 UTC m=+39.487232291 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.822285 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.822334 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.822348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.822368 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.822382 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.925077 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.925164 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.925184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.925206 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.925224 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:26Z","lastTransitionTime":"2025-12-03T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.976218 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.976335 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:26 crc kubenswrapper[4677]: I1203 13:47:26.976338 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:26 crc kubenswrapper[4677]: E1203 13:47:26.976459 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:26 crc kubenswrapper[4677]: E1203 13:47:26.976668 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:26 crc kubenswrapper[4677]: E1203 13:47:26.976790 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.028533 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.028583 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.028598 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.028627 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.028639 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.131405 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.131442 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.131451 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.131467 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.131478 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.234377 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.234498 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.234516 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.234538 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.234557 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.337353 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.337419 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.337439 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.337463 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.337480 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.400573 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.424567 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.439341 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.439827 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.439884 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.439902 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.439926 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.439943 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.454762 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.469813 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.492783 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.515104 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.530705 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.543395 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.543439 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.543451 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.543470 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.543485 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.558440 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.578694 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.593158 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.614248 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.633540 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.646545 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.646576 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.646584 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.646598 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.646607 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.655290 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.676721 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.694559 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.709515 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.724524 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:27Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.750017 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.750065 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.750075 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.750090 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.750102 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.852126 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.852163 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.852171 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.852184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.852193 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.955207 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.955246 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.955258 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.955274 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.955285 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:27Z","lastTransitionTime":"2025-12-03T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:27 crc kubenswrapper[4677]: I1203 13:47:27.975854 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:27 crc kubenswrapper[4677]: E1203 13:47:27.976057 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.057719 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.057789 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.057808 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.057833 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.057850 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.161014 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.161098 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.161129 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.161160 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.161183 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.264021 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.264079 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.264097 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.264117 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.264131 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.366793 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.366832 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.366843 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.366879 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.366892 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.469891 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.470009 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.470037 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.470066 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.470091 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.572855 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.572904 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.572920 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.572939 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.572974 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.676639 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.676681 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.676689 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.676703 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.676713 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.763854 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:28 crc kubenswrapper[4677]: E1203 13:47:28.764218 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:28 crc kubenswrapper[4677]: E1203 13:47:28.764386 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:32.76430995 +0000 UTC m=+43.510642445 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.780762 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.780825 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.780843 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.780867 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.780885 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.883788 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.883855 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.883871 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.883891 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.883906 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.975423 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.975508 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.975546 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:28 crc kubenswrapper[4677]: E1203 13:47:28.975705 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:28 crc kubenswrapper[4677]: E1203 13:47:28.975813 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:28 crc kubenswrapper[4677]: E1203 13:47:28.975920 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.986509 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.986583 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.986603 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.986628 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:28 crc kubenswrapper[4677]: I1203 13:47:28.986645 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:28Z","lastTransitionTime":"2025-12-03T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.089568 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.089668 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.089687 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.089713 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.089731 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.191820 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.191868 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.191884 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.191904 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.191919 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.295006 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.295069 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.295088 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.295116 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.295135 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.398247 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.398338 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.398373 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.398403 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.398426 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.501231 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.501280 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.501289 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.501306 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.501316 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.603293 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.603385 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.603403 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.603423 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.603437 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.706529 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.706575 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.706599 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.706624 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.706641 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.809291 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.809339 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.809351 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.809372 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.809384 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.911662 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.911707 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.911718 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.911734 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.911746 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:29Z","lastTransitionTime":"2025-12-03T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.975596 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:29 crc kubenswrapper[4677]: E1203 13:47:29.975726 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:29 crc kubenswrapper[4677]: I1203 13:47:29.992239 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:29Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.009729 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.013907 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.013959 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.013972 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.013988 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.014001 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.023312 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.038181 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.052990 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.067935 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.082172 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.092621 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.103422 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.114099 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.116099 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.116134 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.116144 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.116161 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.116171 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.130359 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.143011 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.157858 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.169749 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.198456 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5f5b8a90b0993d5593cf7a2ef44136769237aafb6e7991931956cf56815a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"message\\\":\\\"y.go:160\\\\nI1203 13:47:20.926246 5933 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926068 5933 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926596 5933 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 13:47:20.926852 5933 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 13:47:20.927185 5933 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 13:47:20.927196 5933 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 13:47:20.927218 5933 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 13:47:20.927231 5933 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 13:47:20.927236 5933 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 13:47:20.927264 5933 factory.go:656] Stopping watch factory\\\\nI1203 13:47:20.927297 5933 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 13:47:20.927305 5933 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:20.927306 5933 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.218644 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.219061 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.219111 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.219121 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.219135 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.219144 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.230229 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:30Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.321362 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.321411 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.321425 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.321441 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.321452 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.423164 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.423201 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.423209 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.423220 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.423228 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.525637 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.525673 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.525683 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.525695 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.525704 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.628296 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.628388 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.628413 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.628444 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.628466 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.730862 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.730936 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.730993 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.731039 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.731063 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.833775 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.833826 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.833840 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.833862 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.833876 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.936664 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.936716 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.936740 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.936764 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.936779 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:30Z","lastTransitionTime":"2025-12-03T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.975351 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.975397 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:30 crc kubenswrapper[4677]: I1203 13:47:30.975425 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:30 crc kubenswrapper[4677]: E1203 13:47:30.975597 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:30 crc kubenswrapper[4677]: E1203 13:47:30.975685 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:30 crc kubenswrapper[4677]: E1203 13:47:30.975831 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.038880 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.038982 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.039014 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.039032 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.039064 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.146433 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.146520 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.146553 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.146571 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.146580 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.249861 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.250004 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.250029 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.250057 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.250129 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.353035 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.353076 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.353086 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.353101 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.353112 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.456246 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.456323 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.456341 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.456377 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.456398 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.564090 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.564130 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.564143 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.564161 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.564172 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.666897 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.666938 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.666976 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.666999 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.667010 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.768809 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.768854 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.768862 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.768874 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.768882 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.871256 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.871322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.871348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.871378 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.871401 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.975100 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.975168 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.975178 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.975195 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.975208 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:31Z","lastTransitionTime":"2025-12-03T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:31 crc kubenswrapper[4677]: I1203 13:47:31.975487 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:31 crc kubenswrapper[4677]: E1203 13:47:31.975626 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.077846 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.077922 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.077937 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.077987 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.078002 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.181432 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.181826 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.182044 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.182247 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.182440 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.285184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.285299 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.285310 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.285323 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.285330 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.387638 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.387767 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.387801 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.387834 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.387855 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.491001 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.491070 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.491110 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.491143 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.491165 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.594170 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.594232 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.594249 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.594324 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.594343 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.697351 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.697390 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.697399 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.697413 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.697421 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.800280 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.800332 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.800346 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.800368 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.800398 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.808085 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:32 crc kubenswrapper[4677]: E1203 13:47:32.808272 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:32 crc kubenswrapper[4677]: E1203 13:47:32.808334 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:40.808319315 +0000 UTC m=+51.554651780 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.903667 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.903737 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.903750 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.903770 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.903782 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:32Z","lastTransitionTime":"2025-12-03T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.975801 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.975857 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:32 crc kubenswrapper[4677]: E1203 13:47:32.975917 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:32 crc kubenswrapper[4677]: E1203 13:47:32.976025 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:32 crc kubenswrapper[4677]: I1203 13:47:32.975798 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:32 crc kubenswrapper[4677]: E1203 13:47:32.976197 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.006557 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.006604 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.006613 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.006627 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.006638 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.109450 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.109488 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.109495 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.109509 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.109517 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.212726 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.212791 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.212810 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.212834 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.212853 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.253447 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.253537 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.253565 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.253607 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.253639 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: E1203 13:47:33.279313 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.283849 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.283891 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.283899 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.283913 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.283922 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: E1203 13:47:33.296865 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.300122 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.300165 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.300180 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.300200 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.300216 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: E1203 13:47:33.312228 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.316093 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.316125 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.316134 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.316147 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.316155 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: E1203 13:47:33.328705 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.332706 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.332757 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.332769 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.332785 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.332802 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: E1203 13:47:33.351635 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:33Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:33 crc kubenswrapper[4677]: E1203 13:47:33.351749 4677 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.353581 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.353610 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.353618 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.353630 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.353641 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.455918 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.455989 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.456001 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.456016 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.456027 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.559288 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.559341 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.559350 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.559363 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.559372 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.707175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.707220 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.707229 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.707243 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.707251 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.810541 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.810585 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.810600 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.810616 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.810628 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.912921 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.912982 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.912993 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.913007 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.913018 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:33Z","lastTransitionTime":"2025-12-03T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:33 crc kubenswrapper[4677]: I1203 13:47:33.975719 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:33 crc kubenswrapper[4677]: E1203 13:47:33.976004 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.016031 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.016089 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.016108 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.016131 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.016147 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.119333 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.119407 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.119430 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.119457 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.119477 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.222293 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.222373 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.222384 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.222398 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.222410 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.324507 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.324543 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.324554 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.324570 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.324582 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.427269 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.427318 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.427331 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.427350 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.427362 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.529876 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.530138 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.530280 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.530376 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.530473 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.633356 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.633650 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.633745 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.633901 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.634089 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.737241 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.737289 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.737303 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.737321 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.737333 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.839705 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.839772 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.839789 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.839814 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.839831 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.942531 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.942612 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.942635 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.942663 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.942682 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:34Z","lastTransitionTime":"2025-12-03T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.975814 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.975902 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:34 crc kubenswrapper[4677]: E1203 13:47:34.976051 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:34 crc kubenswrapper[4677]: I1203 13:47:34.976120 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:34 crc kubenswrapper[4677]: E1203 13:47:34.976212 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:34 crc kubenswrapper[4677]: E1203 13:47:34.976295 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.045050 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.045131 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.045143 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.045158 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.045169 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.148309 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.148384 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.148399 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.148423 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.148447 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.251559 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.251616 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.251633 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.251655 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.251675 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.320595 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.321998 4677 scope.go:117] "RemoveContainer" containerID="008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.354900 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.355427 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.355437 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.355453 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.355464 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.357224 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.378444 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.396012 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.423723 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.437295 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.451078 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.461296 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.461334 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.461349 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.461369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.461384 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.463423 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.473269 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.487755 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.499047 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.515408 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.528772 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.540843 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.559184 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.564294 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.564336 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.564346 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.564363 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.564374 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.581629 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.604029 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.620538 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:35Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.668034 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.668069 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.668081 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.668096 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.668106 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.770617 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.770670 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.770698 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.770710 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.770719 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.873664 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.873693 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.873701 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.873713 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.873723 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.975324 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:35 crc kubenswrapper[4677]: E1203 13:47:35.975484 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.976054 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.976121 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.976132 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.976147 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:35 crc kubenswrapper[4677]: I1203 13:47:35.976159 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:35Z","lastTransitionTime":"2025-12-03T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.078880 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.078993 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.079015 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.079038 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.079054 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.181633 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.181682 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.181697 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.181718 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.181732 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.285088 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.285158 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.285175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.285199 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.285217 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.389306 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.389412 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.389432 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.389512 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.389546 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.493068 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.493138 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.493162 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.493191 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.493215 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.595531 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.595577 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.595588 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.595606 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.595619 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.698319 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.698348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.698356 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.698369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.698378 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.800387 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.800430 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.800439 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.800455 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.800474 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.903645 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.903705 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.903715 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.903734 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.903744 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:36Z","lastTransitionTime":"2025-12-03T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.975484 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.975540 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:36 crc kubenswrapper[4677]: E1203 13:47:36.975630 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:36 crc kubenswrapper[4677]: E1203 13:47:36.975752 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:36 crc kubenswrapper[4677]: I1203 13:47:36.975821 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:36 crc kubenswrapper[4677]: E1203 13:47:36.975934 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.008146 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.008202 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.008216 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.008238 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.008255 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.110389 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.110437 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.110451 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.110471 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.110485 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.212440 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.212484 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.212495 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.212512 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.212524 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.315449 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.315516 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.315535 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.315562 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.315582 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.418441 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.418528 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.418551 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.418577 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.418594 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.444509 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/1.log" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.447429 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.447764 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.481782 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.514683 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.521481 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.521576 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.521604 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.521636 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.521659 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.543447 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.574821 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.596399 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.617800 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.624376 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.624434 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.624450 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.624477 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.624494 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.637348 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.658210 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.682653 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.697475 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.718239 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.727802 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.727863 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.727883 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.727912 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.727935 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.736764 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.752698 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.766565 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.782271 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.795647 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.808291 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.830337 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.830419 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.830443 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.830475 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.830499 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.933727 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.933811 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.933829 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.933854 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.933871 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:37Z","lastTransitionTime":"2025-12-03T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:37 crc kubenswrapper[4677]: I1203 13:47:37.975133 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:37 crc kubenswrapper[4677]: E1203 13:47:37.975263 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.037107 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.037165 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.037183 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.037204 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.037221 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.139477 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.139514 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.139523 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.139536 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.139544 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.242434 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.242495 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.242513 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.242542 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.242559 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.345671 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.345726 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.345740 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.345758 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.345771 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.448728 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.448798 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.448816 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.448839 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.448857 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.458818 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/2.log" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.460179 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/1.log" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.464024 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53" exitCode=1 Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.464078 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.464139 4677 scope.go:117] "RemoveContainer" containerID="008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.465382 4677 scope.go:117] "RemoveContainer" containerID="1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53" Dec 03 13:47:38 crc kubenswrapper[4677]: E1203 13:47:38.467060 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.484559 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.503787 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.527043 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.548109 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.551367 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.551437 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.551459 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.551513 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.551538 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.565667 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.582693 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.598528 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.611442 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.626425 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.648824 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008de23584019f7f4bec50fbdbebf66bcea3aeda7db081d87821fef4cc3bcc9b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"message\\\":\\\"4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 13:47:22.238424 6101 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 13:47:22.238426 6101 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 13:47:22.238435 6101 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 13:47:22.238437 6101 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1203 13:47:22.238443 6101 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 13:47:22.238435 6101 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nI1203 13:47:22.238454 6101 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-controller-manager-operator for network=default : 945.837µs\\\\nI1203 13:47:22.238418 6101 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 765.102µs\\\\nI1203 13:47:22.238420 6101 handler.go:208] Removed *v1.Node event handler 2\\\\nF1203 13:47:22.238477 6101 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.653860 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.653913 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.653924 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.653942 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.653976 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.671008 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.686278 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.699125 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.716498 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.731745 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.745268 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.756582 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.756621 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.756630 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.756643 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.756652 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.759198 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:38Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.858904 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.858978 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.858991 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.859007 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.859018 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.962159 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.962651 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.962712 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.962748 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.962794 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:38Z","lastTransitionTime":"2025-12-03T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.976041 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.976162 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:38 crc kubenswrapper[4677]: E1203 13:47:38.976204 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:38 crc kubenswrapper[4677]: E1203 13:47:38.976302 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:38 crc kubenswrapper[4677]: I1203 13:47:38.976351 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:38 crc kubenswrapper[4677]: E1203 13:47:38.976394 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.069365 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.069423 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.069434 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.069451 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.069464 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.172656 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.173139 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.173281 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.173405 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.173521 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.276134 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.276203 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.276219 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.276241 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.276257 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.379145 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.379208 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.379225 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.379249 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.379267 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.472109 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/2.log" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.477456 4677 scope.go:117] "RemoveContainer" containerID="1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53" Dec 03 13:47:39 crc kubenswrapper[4677]: E1203 13:47:39.477829 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.484322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.484382 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.484399 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.484422 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.484440 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.498055 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.518842 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.533359 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.547031 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.560090 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.579560 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.587254 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.587452 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.587525 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.587594 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.587656 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.592350 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.606894 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.620124 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.640925 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.663420 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.677344 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.690765 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.690815 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.690995 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.691008 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.691026 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.691038 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.705610 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.717172 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.730115 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.742522 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.793066 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.793116 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.793128 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.793144 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.793156 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.896137 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.896177 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.896188 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.896203 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.896214 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:39Z","lastTransitionTime":"2025-12-03T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.975852 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:39 crc kubenswrapper[4677]: E1203 13:47:39.976212 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:39 crc kubenswrapper[4677]: I1203 13:47:39.989424 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:39Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.000479 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.000527 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.000542 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.000564 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.000581 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.007253 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.023224 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.041530 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.055750 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.083727 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.103159 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.103199 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.103207 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.103221 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.103229 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.107406 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.122515 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.139503 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.161848 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.173608 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.189525 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.203332 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.205646 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.205696 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.205710 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.205729 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.205743 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.215621 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.228318 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.242785 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.252792 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:40Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.307640 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.307703 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.307727 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.307755 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.307777 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.410526 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.410629 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.410647 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.410674 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.410692 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.514326 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.514392 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.514413 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.514445 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.514470 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.617264 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.617306 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.617315 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.617329 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.617338 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.720342 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.720385 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.720396 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.720411 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.720423 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.796289 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.796548 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:48:12.796500568 +0000 UTC m=+83.542833073 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.796718 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.796892 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.797009 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:48:12.796985272 +0000 UTC m=+83.543317747 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.823223 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.823267 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.823278 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.823293 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.823305 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.897549 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.897692 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.897757 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.897873 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:47:56.897843062 +0000 UTC m=+67.644175557 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.897926 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.897993 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.897780 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898005 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898195 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898115 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:48:12.898088329 +0000 UTC m=+83.644420834 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.898263 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898333 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:48:12.898301625 +0000 UTC m=+83.644634080 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898430 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898459 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898483 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.898551 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:48:12.898528652 +0000 UTC m=+83.644861257 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.925420 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.925468 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.925477 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.925501 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.925520 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:40Z","lastTransitionTime":"2025-12-03T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.975478 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.975483 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:40 crc kubenswrapper[4677]: I1203 13:47:40.975521 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.976069 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.976233 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:40 crc kubenswrapper[4677]: E1203 13:47:40.976399 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.028867 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.028921 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.028935 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.028981 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.029000 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.132263 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.132311 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.132322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.132339 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.132351 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.235304 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.235367 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.235377 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.235394 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.235405 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.337486 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.337537 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.337547 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.337564 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.337574 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.440103 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.440148 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.440159 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.440176 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.440189 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.542415 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.542440 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.542448 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.542460 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.542468 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.644641 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.644916 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.645005 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.645080 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.645141 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.748308 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.748596 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.748605 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.748619 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.748630 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.852511 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.852596 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.852623 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.852651 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.852673 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.955763 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.955814 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.955830 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.955851 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.955867 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:41Z","lastTransitionTime":"2025-12-03T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:41 crc kubenswrapper[4677]: I1203 13:47:41.975163 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:41 crc kubenswrapper[4677]: E1203 13:47:41.975294 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.058311 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.058363 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.058382 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.058405 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.058423 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.161787 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.161844 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.161861 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.161886 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.161904 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.264357 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.264395 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.264406 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.264420 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.264430 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.367636 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.367685 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.367698 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.367718 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.367734 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.471428 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.471509 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.471551 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.471589 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.471613 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.575493 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.575547 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.575559 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.575576 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.575588 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.678643 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.678738 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.678756 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.678783 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.678800 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.781267 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.781316 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.781331 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.781348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.781362 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.883765 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.883824 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.883850 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.883876 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.883892 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.975364 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.975408 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.975444 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:42 crc kubenswrapper[4677]: E1203 13:47:42.975505 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:42 crc kubenswrapper[4677]: E1203 13:47:42.975577 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:42 crc kubenswrapper[4677]: E1203 13:47:42.975629 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.986575 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.986621 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.986632 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.986647 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:42 crc kubenswrapper[4677]: I1203 13:47:42.986660 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:42Z","lastTransitionTime":"2025-12-03T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.088741 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.088790 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.088803 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.088817 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.088825 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.191193 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.191236 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.191246 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.191263 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.191274 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.294216 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.294259 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.294268 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.294283 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.294293 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.397748 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.397825 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.397843 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.397868 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.397892 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.500819 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.500874 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.500885 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.500902 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.500916 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.603783 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.603825 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.603833 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.603847 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.603858 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.681142 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.681191 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.681200 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.681214 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.681226 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: E1203 13:47:43.693879 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.701704 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.701773 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.701784 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.701800 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.701813 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: E1203 13:47:43.719704 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.724996 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.725037 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.725046 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.725062 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.725072 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: E1203 13:47:43.740902 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.745537 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.745600 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.745616 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.745636 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.745650 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: E1203 13:47:43.759069 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.763262 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.763288 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.763295 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.763310 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.763319 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: E1203 13:47:43.778297 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:43Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:43 crc kubenswrapper[4677]: E1203 13:47:43.778791 4677 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.780844 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.780899 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.780919 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.780943 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.780981 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.885007 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.885051 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.885061 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.885077 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.885091 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.976198 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:43 crc kubenswrapper[4677]: E1203 13:47:43.976407 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.987653 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.987698 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.987713 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.987735 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:43 crc kubenswrapper[4677]: I1203 13:47:43.987751 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:43Z","lastTransitionTime":"2025-12-03T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.090551 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.090637 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.090672 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.090699 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.090716 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.193729 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.193762 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.193771 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.193786 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.193796 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.296773 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.296860 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.296900 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.296932 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.296997 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.400339 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.400411 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.400433 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.400462 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.400487 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.485567 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.499782 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.501661 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.503727 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.503756 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.503765 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.503777 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.503786 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.517904 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.532076 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.547721 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.566530 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.579841 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.594871 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.606644 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.606703 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.606717 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.606737 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.606752 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.612199 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.624682 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.641791 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.660654 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.672936 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.685747 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.700042 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.708786 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.708835 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.708850 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.708871 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.708887 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.721666 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.730478 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.740463 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:44Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.811205 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.811255 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.811268 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.811287 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.811299 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.914027 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.914114 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.914137 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.914167 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.914224 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:44Z","lastTransitionTime":"2025-12-03T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.976052 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.976118 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:44 crc kubenswrapper[4677]: I1203 13:47:44.976232 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:44 crc kubenswrapper[4677]: E1203 13:47:44.976250 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:44 crc kubenswrapper[4677]: E1203 13:47:44.976339 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:44 crc kubenswrapper[4677]: E1203 13:47:44.976443 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.016630 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.016674 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.016700 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.016722 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.016734 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.119175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.119226 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.119240 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.119259 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.119272 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.222097 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.222172 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.222196 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.222226 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.222248 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.325552 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.325631 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.325655 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.325683 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.325705 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.429132 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.429212 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.429234 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.429265 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.429286 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.532706 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.532811 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.532832 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.532860 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.532879 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.636272 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.636337 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.636355 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.636381 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.636402 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.740701 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.740763 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.740781 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.740805 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.740825 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.843988 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.844034 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.844044 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.844059 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.844067 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.946591 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.946676 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.946695 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.946717 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.946764 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:45Z","lastTransitionTime":"2025-12-03T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:45 crc kubenswrapper[4677]: I1203 13:47:45.975293 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:45 crc kubenswrapper[4677]: E1203 13:47:45.975573 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.049814 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.049859 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.049877 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.049894 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.049906 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.152636 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.152666 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.152674 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.152687 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.152696 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.255421 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.255459 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.255470 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.255484 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.255494 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.357175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.357213 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.357221 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.357235 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.357245 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.459341 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.459369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.459379 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.459391 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.459399 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.562047 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.562403 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.562538 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.562668 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.562785 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.666328 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.666665 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.666792 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.666908 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.667005 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.770395 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.770732 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.771050 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.771211 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.771337 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.874131 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.874194 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.874218 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.874248 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.874268 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.975214 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.975290 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.975290 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:46 crc kubenswrapper[4677]: E1203 13:47:46.975348 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:46 crc kubenswrapper[4677]: E1203 13:47:46.975437 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:46 crc kubenswrapper[4677]: E1203 13:47:46.975544 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.976473 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.976567 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.976629 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.976654 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:46 crc kubenswrapper[4677]: I1203 13:47:46.976716 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:46Z","lastTransitionTime":"2025-12-03T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.080744 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.080840 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.080881 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.080921 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.080991 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.184283 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.184335 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.184348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.184365 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.184379 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.287502 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.287555 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.287566 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.287584 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.287595 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.389429 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.389474 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.389484 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.389498 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.389509 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.492054 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.492130 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.492142 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.492164 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.492179 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.595528 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.595634 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.595665 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.595699 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.595722 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.698682 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.698765 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.698786 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.698821 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.698846 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.801895 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.802031 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.802050 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.802077 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.802096 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.905160 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.905224 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.905247 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.905451 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.905571 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:47Z","lastTransitionTime":"2025-12-03T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:47 crc kubenswrapper[4677]: I1203 13:47:47.976064 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:47 crc kubenswrapper[4677]: E1203 13:47:47.976278 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.007703 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.007974 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.007988 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.008008 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.008021 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.111925 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.112042 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.112065 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.112097 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.112121 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.214904 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.215033 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.215061 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.215097 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.215124 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.318716 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.318781 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.318800 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.318826 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.318843 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.421154 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.421237 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.421259 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.421293 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.421316 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.525559 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.525648 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.525684 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.525720 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.525748 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.629342 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.629386 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.629395 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.629410 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.629421 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.732393 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.732435 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.732445 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.732461 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.732472 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.836123 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.836212 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.836255 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.836288 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.836314 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.939803 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.939866 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.939885 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.939905 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.939917 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:48Z","lastTransitionTime":"2025-12-03T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.975541 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.975545 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:48 crc kubenswrapper[4677]: I1203 13:47:48.975571 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:48 crc kubenswrapper[4677]: E1203 13:47:48.975831 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:48 crc kubenswrapper[4677]: E1203 13:47:48.975925 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:48 crc kubenswrapper[4677]: E1203 13:47:48.975693 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.043146 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.043214 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.043224 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.043430 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.043443 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.147297 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.147346 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.147360 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.147383 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.147398 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.250345 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.250398 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.250409 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.250426 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.250436 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.353107 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.353189 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.353205 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.353228 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.353243 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.456215 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.456245 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.456255 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.456267 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.456275 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.559479 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.559523 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.559532 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.559548 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.559557 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.663632 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.663696 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.663709 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.663729 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.663744 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.766330 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.766803 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.766892 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.767000 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.767078 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.869849 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.870223 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.870328 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.870437 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.870519 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.972166 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.972204 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.972213 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.972227 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.972237 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:49Z","lastTransitionTime":"2025-12-03T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.975065 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:49 crc kubenswrapper[4677]: E1203 13:47:49.975183 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:49 crc kubenswrapper[4677]: I1203 13:47:49.991376 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:49Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.002467 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.012826 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.025414 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.037508 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.061245 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.073837 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.073915 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.073931 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.073977 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.073993 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.077181 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.105958 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.125906 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.139534 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.150532 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.162976 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.173641 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.176298 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.176322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.176329 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.176342 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.176350 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.190857 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.209722 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.219565 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.231220 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.247825 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:50Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.278145 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.278175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.278185 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.278197 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.278206 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.380989 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.381646 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.381892 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.382167 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.382384 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.486316 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.486359 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.486370 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.486387 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.486399 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.590315 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.590377 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.590400 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.590429 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.590452 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.693782 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.693821 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.693833 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.693850 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.693862 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.796209 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.796247 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.796255 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.796270 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.796281 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.898308 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.898348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.898355 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.898369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.898378 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:50Z","lastTransitionTime":"2025-12-03T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.975322 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.975403 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:50 crc kubenswrapper[4677]: E1203 13:47:50.975533 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:50 crc kubenswrapper[4677]: I1203 13:47:50.975637 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:50 crc kubenswrapper[4677]: E1203 13:47:50.975719 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:50 crc kubenswrapper[4677]: E1203 13:47:50.976103 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.000867 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.000924 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.000984 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.001012 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.001033 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.104555 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.104633 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.104645 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.104672 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.104686 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.207444 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.207513 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.207532 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.207557 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.207574 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.311705 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.311804 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.311819 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.311842 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.311868 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.419977 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.420050 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.420069 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.420095 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.420112 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.522533 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.522594 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.522611 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.522633 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.522649 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.625700 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.625749 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.625762 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.625779 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.625791 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.727938 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.727999 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.728012 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.728027 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.728039 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.830134 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.830183 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.830196 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.830213 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.830226 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.932501 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.932538 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.932548 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.932560 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.932568 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:51Z","lastTransitionTime":"2025-12-03T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.976059 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:51 crc kubenswrapper[4677]: E1203 13:47:51.976193 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:51 crc kubenswrapper[4677]: I1203 13:47:51.976896 4677 scope.go:117] "RemoveContainer" containerID="1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53" Dec 03 13:47:51 crc kubenswrapper[4677]: E1203 13:47:51.977084 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.035408 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.035454 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.035465 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.035482 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.035496 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.137574 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.137650 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.137671 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.137697 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.137716 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.239929 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.240003 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.240020 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.240042 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.240058 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.342463 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.342857 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.342876 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.342900 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.342920 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.446719 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.446799 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.446812 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.446840 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.446864 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.550179 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.550218 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.550226 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.550240 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.550249 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.653374 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.653493 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.653521 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.653550 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.653570 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.756823 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.756906 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.756925 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.756983 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.757007 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.859730 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.859798 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.859824 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.859853 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.859872 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.962982 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.963072 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.963085 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.963139 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.963152 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:52Z","lastTransitionTime":"2025-12-03T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.975706 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.975782 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.975735 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:52 crc kubenswrapper[4677]: E1203 13:47:52.975887 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:52 crc kubenswrapper[4677]: E1203 13:47:52.976052 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:52 crc kubenswrapper[4677]: E1203 13:47:52.976233 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:52 crc kubenswrapper[4677]: I1203 13:47:52.989552 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.066406 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.066466 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.066479 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.066497 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.066510 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.170376 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.170458 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.170480 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.170509 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.170531 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.273276 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.273336 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.273351 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.273369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.273385 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.376357 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.376427 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.376446 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.376472 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.376491 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.479138 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.479196 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.479215 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.479238 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.479257 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.582526 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.582597 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.582679 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.582718 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.582743 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.686350 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.686404 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.686417 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.686433 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.686443 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.789238 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.789278 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.789289 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.789306 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.789317 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.891880 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.891921 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.891932 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.891970 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.891982 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.908049 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.908094 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.908105 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.908120 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.908133 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: E1203 13:47:53.925731 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:53Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.929559 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.929586 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.929595 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.929607 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.929616 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: E1203 13:47:53.941012 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:53Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.944752 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.944786 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.944797 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.944813 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.944825 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: E1203 13:47:53.956769 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:53Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.960427 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.960476 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.960488 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.960505 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.960516 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.976162 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:53 crc kubenswrapper[4677]: E1203 13:47:53.976299 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:53 crc kubenswrapper[4677]: E1203 13:47:53.976288 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:53Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.980775 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.980809 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.980819 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.980833 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.980844 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:53 crc kubenswrapper[4677]: E1203 13:47:53.992849 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:53Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:53 crc kubenswrapper[4677]: E1203 13:47:53.993018 4677 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.994992 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.995021 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.995032 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.995049 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:53 crc kubenswrapper[4677]: I1203 13:47:53.995062 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:53Z","lastTransitionTime":"2025-12-03T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.098211 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.098244 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.098253 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.098267 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.098277 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.201090 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.201136 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.201149 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.201166 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.201177 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.303173 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.303205 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.303215 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.303229 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.303241 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.405051 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.405093 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.405104 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.405120 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.405130 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.508153 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.508205 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.508232 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.508249 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.508259 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.611087 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.611122 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.611131 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.611146 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.611156 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.713431 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.713463 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.713471 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.713485 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.713494 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.818438 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.818488 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.818500 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.818516 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.818529 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.920360 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.920390 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.920398 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.920412 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.920421 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:54Z","lastTransitionTime":"2025-12-03T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.975064 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:54 crc kubenswrapper[4677]: E1203 13:47:54.975223 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.975276 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:54 crc kubenswrapper[4677]: I1203 13:47:54.975284 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:54 crc kubenswrapper[4677]: E1203 13:47:54.975427 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:54 crc kubenswrapper[4677]: E1203 13:47:54.975548 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.023245 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.023291 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.023301 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.023316 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.023325 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.125262 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.125310 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.125322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.125338 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.125351 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.227941 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.227994 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.228006 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.228022 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.228033 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.329696 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.329735 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.329743 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.329755 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.329764 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.432030 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.432072 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.432082 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.432097 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.432108 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.535893 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.535942 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.535970 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.535987 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.535998 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.644123 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.644157 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.644165 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.644178 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.644186 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.746138 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.746188 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.746200 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.746218 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.746232 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.848670 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.848741 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.848762 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.848785 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.848803 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.951892 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.951944 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.951974 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.951992 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.952005 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:55Z","lastTransitionTime":"2025-12-03T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:55 crc kubenswrapper[4677]: I1203 13:47:55.975822 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:55 crc kubenswrapper[4677]: E1203 13:47:55.976017 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.055020 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.055065 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.055075 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.055093 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.055105 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.158159 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.158248 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.158272 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.158314 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.158344 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.260412 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.260451 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.260467 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.260481 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.260490 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.362754 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.362795 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.362803 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.362818 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.362827 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.465625 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.465668 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.465679 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.465694 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.465707 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.567941 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.568001 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.568013 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.568028 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.568041 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.670402 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.670428 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.670435 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.670450 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.670460 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.772840 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.772872 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.772883 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.772898 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.772909 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.875184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.875215 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.875224 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.875237 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.875246 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.971573 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:56 crc kubenswrapper[4677]: E1203 13:47:56.971700 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:56 crc kubenswrapper[4677]: E1203 13:47:56.971746 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:48:28.971731559 +0000 UTC m=+99.718064004 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.975514 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.975593 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.975523 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:56 crc kubenswrapper[4677]: E1203 13:47:56.975654 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:56 crc kubenswrapper[4677]: E1203 13:47:56.976240 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:56 crc kubenswrapper[4677]: E1203 13:47:56.976322 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.979111 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.979140 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.979148 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.979161 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:56 crc kubenswrapper[4677]: I1203 13:47:56.979171 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:56Z","lastTransitionTime":"2025-12-03T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.080967 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.081003 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.081011 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.081028 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.081038 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.183329 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.183383 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.183395 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.183412 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.183426 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.286282 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.286321 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.286330 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.286342 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.286351 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.389270 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.389301 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.389310 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.389322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.389331 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.491738 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.491772 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.491780 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.491794 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.491805 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.594460 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.594493 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.594506 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.594520 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.594531 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.696723 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.696759 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.696768 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.696782 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.696792 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.798570 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.798611 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.798623 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.798639 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.798651 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.903903 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.903942 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.903977 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.903994 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.904008 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:57Z","lastTransitionTime":"2025-12-03T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:57 crc kubenswrapper[4677]: I1203 13:47:57.975875 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:57 crc kubenswrapper[4677]: E1203 13:47:57.976013 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.005920 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.005982 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.005993 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.006008 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.006020 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.108584 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.108615 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.108623 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.108636 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.108644 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.210900 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.211007 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.211032 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.211063 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.211085 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.314455 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.314500 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.314522 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.314540 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.314553 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.416600 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.416647 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.416662 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.416680 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.416694 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.519276 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.519325 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.519337 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.519355 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.519367 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.621566 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.621603 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.621612 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.621627 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.621638 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.723858 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.723888 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.723899 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.723915 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.723926 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.826291 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.826368 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.826390 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.826418 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.826437 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.929648 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.929684 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.929693 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.929706 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.929716 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:58Z","lastTransitionTime":"2025-12-03T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.976078 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.976120 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:47:58 crc kubenswrapper[4677]: I1203 13:47:58.976193 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:47:58 crc kubenswrapper[4677]: E1203 13:47:58.976201 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:47:58 crc kubenswrapper[4677]: E1203 13:47:58.976413 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:47:58 crc kubenswrapper[4677]: E1203 13:47:58.976471 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.032701 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.032829 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.033292 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.033386 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.033693 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.136903 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.136983 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.137000 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.137024 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.137044 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.239346 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.239383 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.239397 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.239425 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.239440 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.341845 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.341913 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.341936 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.342032 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.342061 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.444701 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.444743 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.444754 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.444770 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.444783 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.547211 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.547249 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.547260 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.547274 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.547286 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.548918 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/0.log" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.548974 4677 generic.go:334] "Generic (PLEG): container finished" podID="511a6f65-7dac-4f37-a15e-3a24339f80f4" containerID="0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19" exitCode=1 Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.549001 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerDied","Data":"0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.549367 4677 scope.go:117] "RemoveContainer" containerID="0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.564726 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545df6df-eecf-4d8b-8981-0b57b96b227a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.577109 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.587288 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.599567 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:58Z\\\",\\\"message\\\":\\\"2025-12-03T13:47:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54\\\\n2025-12-03T13:47:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54 to /host/opt/cni/bin/\\\\n2025-12-03T13:47:13Z [verbose] multus-daemon started\\\\n2025-12-03T13:47:13Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:47:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.614693 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.625231 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.637700 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.648769 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.649192 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.649227 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.649236 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.649250 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.649259 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.658960 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.671528 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.684112 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.700034 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.712659 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.722994 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.746273 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.750757 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.750792 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.750801 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.750815 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.750824 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.758481 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.769150 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.779627 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.796338 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.853679 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.853719 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.853728 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.853745 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.853758 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.955520 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.955564 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.955576 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.955594 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.955608 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:47:59Z","lastTransitionTime":"2025-12-03T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.975930 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:47:59 crc kubenswrapper[4677]: E1203 13:47:59.976169 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:47:59 crc kubenswrapper[4677]: I1203 13:47:59.997086 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:59Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.012398 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.028695 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.042263 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.057896 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.057927 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.057935 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.057961 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.057970 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.063960 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.079219 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.091122 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.101688 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545df6df-eecf-4d8b-8981-0b57b96b227a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.115284 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.127964 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.141679 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:58Z\\\",\\\"message\\\":\\\"2025-12-03T13:47:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54\\\\n2025-12-03T13:47:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54 to /host/opt/cni/bin/\\\\n2025-12-03T13:47:13Z [verbose] multus-daemon started\\\\n2025-12-03T13:47:13Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:47:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.156512 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.159849 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.159883 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.159895 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.159909 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.159917 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.170328 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.181880 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.194899 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.206982 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.219259 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.229984 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.240793 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.262545 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.262596 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.262610 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.262627 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.262637 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.364428 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.364471 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.364483 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.364499 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.364510 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.466439 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.466492 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.466507 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.466527 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.466542 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.555677 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/0.log" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.555905 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerStarted","Data":"6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.570841 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.571146 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.571233 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.571313 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.571389 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.581609 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:58Z\\\",\\\"message\\\":\\\"2025-12-03T13:47:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54\\\\n2025-12-03T13:47:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54 to /host/opt/cni/bin/\\\\n2025-12-03T13:47:13Z [verbose] multus-daemon started\\\\n2025-12-03T13:47:13Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:47:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.604684 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.621760 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.641382 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.655741 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545df6df-eecf-4d8b-8981-0b57b96b227a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.674279 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.674315 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.674326 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.674344 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.674454 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.675912 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.688867 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.705205 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.718746 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.728441 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.737518 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.751569 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.764764 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.777181 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.777234 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.777250 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.777271 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.777289 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.778379 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.796908 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.818447 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.839733 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.852504 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.866204 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:00Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.879985 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.880008 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.880016 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.880030 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.880039 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.975305 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.975313 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:00 crc kubenswrapper[4677]: E1203 13:48:00.975452 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:00 crc kubenswrapper[4677]: E1203 13:48:00.975587 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.975831 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:00 crc kubenswrapper[4677]: E1203 13:48:00.976062 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.982060 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.982114 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.982128 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.982142 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:00 crc kubenswrapper[4677]: I1203 13:48:00.982154 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:00Z","lastTransitionTime":"2025-12-03T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.085509 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.085550 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.085561 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.085576 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.085587 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.187798 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.187880 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.187888 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.187903 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.187911 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.290457 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.290502 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.290513 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.290528 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.290539 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.392841 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.392890 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.392903 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.392921 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.392933 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.496711 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.496750 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.496757 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.496769 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.496778 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.598699 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.598741 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.598753 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.598769 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.598781 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.701120 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.701169 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.701182 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.701205 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.701217 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.803759 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.803798 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.803807 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.803822 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.803832 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.906621 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.906664 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.906674 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.906688 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.906697 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:01Z","lastTransitionTime":"2025-12-03T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:01 crc kubenswrapper[4677]: I1203 13:48:01.975686 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:01 crc kubenswrapper[4677]: E1203 13:48:01.975823 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.009278 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.009333 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.009350 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.009372 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.009389 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.112308 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.112342 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.112354 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.112369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.112380 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.214684 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.214733 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.214744 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.214757 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.214766 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.317308 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.317348 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.317356 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.317369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.317378 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.420027 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.420096 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.420109 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.420127 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.420140 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.522378 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.522434 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.522443 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.522460 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.522469 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.624886 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.624963 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.624975 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.624991 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.624999 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.727245 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.727286 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.727294 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.727306 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.727317 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.829982 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.830021 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.830031 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.830045 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.830054 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.932848 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.932902 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.932915 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.932929 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.932944 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:02Z","lastTransitionTime":"2025-12-03T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.975532 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.975601 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:02 crc kubenswrapper[4677]: I1203 13:48:02.975690 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:02 crc kubenswrapper[4677]: E1203 13:48:02.975809 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:02 crc kubenswrapper[4677]: E1203 13:48:02.975931 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:02 crc kubenswrapper[4677]: E1203 13:48:02.976064 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.035369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.035415 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.035423 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.035438 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.035448 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.137610 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.137676 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.137693 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.137716 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.137734 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.240274 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.240318 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.240328 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.240341 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.240350 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.342031 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.342064 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.342072 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.342084 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.342109 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.444881 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.444965 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.444975 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.444989 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.445001 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.548207 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.548264 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.548281 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.548304 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.548320 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.651243 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.651309 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.651335 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.651362 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.651389 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.754751 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.754795 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.754804 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.754836 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.754847 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.856849 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.856913 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.856930 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.856989 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.857015 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.960276 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.960325 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.960342 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.960363 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.960378 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:03Z","lastTransitionTime":"2025-12-03T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:03 crc kubenswrapper[4677]: I1203 13:48:03.976051 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:03 crc kubenswrapper[4677]: E1203 13:48:03.976209 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.063691 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.063772 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.063795 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.063825 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.063846 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.166858 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.166917 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.166930 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.166965 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.166980 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.242864 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.242933 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.243225 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.243330 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.243356 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.266461 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.271275 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.271328 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.271344 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.271367 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.271383 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.290548 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.295250 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.295307 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.295327 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.295350 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.295369 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.309369 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.313765 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.313809 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.313828 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.313851 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.313869 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.326119 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.330417 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.330463 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.330481 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.330504 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.330522 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.349107 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f88142c8-32e7-4ad1-b5c9-79d13ec2e1ea\\\",\\\"systemUUID\\\":\\\"1bc9448c-c189-4fc5-b5df-6e0a45902353\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:04Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.349323 4677 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.352231 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.352299 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.352322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.352355 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.352378 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.455691 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.455762 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.455779 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.455802 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.455820 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.565985 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.566037 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.566054 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.566080 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.566096 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.669577 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.669675 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.669711 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.669741 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.669763 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.772174 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.772254 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.772304 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.772336 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.772359 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.875396 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.875470 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.875493 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.875521 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.875542 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.976333 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.976375 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.976554 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.976600 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.976700 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:04 crc kubenswrapper[4677]: E1203 13:48:04.977387 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.977902 4677 scope.go:117] "RemoveContainer" containerID="1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.978205 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.978280 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.978299 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.978321 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:04 crc kubenswrapper[4677]: I1203 13:48:04.978335 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:04Z","lastTransitionTime":"2025-12-03T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.083085 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.083134 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.083147 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.083168 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.083181 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.186666 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.186726 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.186739 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.186760 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.186773 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.288844 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.288934 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.288967 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.288999 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.289016 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.392290 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.392338 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.392347 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.392364 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.392374 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.495729 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.495786 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.495797 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.495817 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.495831 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.575750 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/2.log" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.584914 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.599845 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.599904 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.599933 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.600009 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.600031 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.703609 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.703689 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.703709 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.703735 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.703752 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.807246 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.807317 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.807340 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.807364 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.807380 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.909617 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.909668 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.909684 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.909704 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.909720 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:05Z","lastTransitionTime":"2025-12-03T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:05 crc kubenswrapper[4677]: I1203 13:48:05.975094 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:05 crc kubenswrapper[4677]: E1203 13:48:05.975266 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.012505 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.012570 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.012585 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.012609 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.012626 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.114524 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.114562 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.114573 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.114588 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.114598 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.218752 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.219049 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.219057 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.219071 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.219079 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.321921 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.322060 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.322081 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.322107 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.322124 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.425488 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.425546 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.425558 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.425576 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.425591 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.528268 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.528317 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.528332 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.528351 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.528361 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.589445 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.604408 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545df6df-eecf-4d8b-8981-0b57b96b227a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.624209 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.630473 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.630535 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.630551 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.630577 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.630592 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.636325 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.654614 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:58Z\\\",\\\"message\\\":\\\"2025-12-03T13:47:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54\\\\n2025-12-03T13:47:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54 to /host/opt/cni/bin/\\\\n2025-12-03T13:47:13Z [verbose] multus-daemon started\\\\n2025-12-03T13:47:13Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:47:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.675679 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.686584 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.697497 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.710688 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.719698 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.731393 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.732806 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.732834 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.732842 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.732856 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.732869 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.742016 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.753590 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.764753 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.774995 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.791821 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.802491 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.813616 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.824678 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.834778 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.834831 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.834844 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.834864 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.834877 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.843155 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:06Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.937036 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.937093 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.937109 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.937135 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.937152 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:06Z","lastTransitionTime":"2025-12-03T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.975481 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.975519 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:06 crc kubenswrapper[4677]: I1203 13:48:06.975481 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:06 crc kubenswrapper[4677]: E1203 13:48:06.975630 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:06 crc kubenswrapper[4677]: E1203 13:48:06.975709 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:06 crc kubenswrapper[4677]: E1203 13:48:06.975775 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.039436 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.039477 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.039488 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.039506 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.039518 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.142015 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.142064 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.142076 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.142094 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.142105 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.245455 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.245494 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.245504 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.245518 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.245528 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.347741 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.347784 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.347796 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.347814 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.347825 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.450122 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.450164 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.450176 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.450193 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.450205 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.552843 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.552878 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.552887 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.552903 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.552911 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.593238 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/3.log" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.593931 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/2.log" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.596352 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" exitCode=1 Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.596380 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.596421 4677 scope.go:117] "RemoveContainer" containerID="1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.597413 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 13:48:07 crc kubenswrapper[4677]: E1203 13:48:07.597672 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.612719 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.628232 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.643027 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.654939 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.655547 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.655597 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.655630 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.655645 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.655655 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.664206 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.685077 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.695535 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.707933 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.718389 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.734486 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe26be8a717b9d1dfcb06add342becc5e4f169bb6f4be87a061b4456b1fac53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:37Z\\\",\\\"message\\\":\\\"ces.lbConfig(nil)\\\\nF1203 13:47:37.381577 6317 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1203 13:47:37.381527 6317 services_controller.go:451] Built service openshift-kube-scheduler/scheduler cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler/scheduler_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:48:07Z\\\",\\\"message\\\":\\\"ultus/network-metrics-daemon-j7h62 openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql openshift-ovn-kubernetes/ovnkube-node-pm7pz]\\\\nI1203 13:48:06.938254 6697 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1203 13:48:06.938274 6697 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938287 6697 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938298 6697 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz in node crc\\\\nI1203 13:48:06.938306 6697 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz after 0 failed attempt(s)\\\\nI1203 13:48:06.938315 6697 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938336 6697 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 13:48:06.938394 6697 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.744478 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545df6df-eecf-4d8b-8981-0b57b96b227a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.755608 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.759438 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.759469 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.759478 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.759493 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.759502 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.772439 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.786976 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:58Z\\\",\\\"message\\\":\\\"2025-12-03T13:47:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54\\\\n2025-12-03T13:47:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54 to /host/opt/cni/bin/\\\\n2025-12-03T13:47:13Z [verbose] multus-daemon started\\\\n2025-12-03T13:47:13Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:47:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.801329 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.810212 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.819825 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.831265 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.840696 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:07Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.861793 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.861829 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.861840 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.861855 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.861865 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.964514 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.964569 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.964580 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.964596 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.964608 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:07Z","lastTransitionTime":"2025-12-03T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:07 crc kubenswrapper[4677]: I1203 13:48:07.976030 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:07 crc kubenswrapper[4677]: E1203 13:48:07.976168 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.067138 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.067211 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.067234 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.067263 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.067288 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.170237 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.170275 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.170285 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.170301 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.170309 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.273932 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.274023 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.274041 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.274065 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.274083 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.377489 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.377548 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.377571 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.377601 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.377624 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.480577 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.480646 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.480662 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.480682 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.480701 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.583211 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.583250 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.583265 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.583284 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.583296 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.603580 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/3.log" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.608091 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 13:48:08 crc kubenswrapper[4677]: E1203 13:48:08.608229 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.623723 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:58Z\\\",\\\"message\\\":\\\"2025-12-03T13:47:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54\\\\n2025-12-03T13:47:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54 to /host/opt/cni/bin/\\\\n2025-12-03T13:47:13Z [verbose] multus-daemon started\\\\n2025-12-03T13:47:13Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:47:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.637373 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.648029 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.660403 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.669353 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545df6df-eecf-4d8b-8981-0b57b96b227a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.683236 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.686115 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.686158 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.686175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.686196 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.686209 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.693973 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.706487 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.716337 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.726410 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.735318 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.748384 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.758358 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.768343 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.777661 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.788602 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.788639 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.788648 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.788688 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.788700 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.796551 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:48:07Z\\\",\\\"message\\\":\\\"ultus/network-metrics-daemon-j7h62 openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql openshift-ovn-kubernetes/ovnkube-node-pm7pz]\\\\nI1203 13:48:06.938254 6697 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1203 13:48:06.938274 6697 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938287 6697 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938298 6697 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz in node crc\\\\nI1203 13:48:06.938306 6697 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz after 0 failed attempt(s)\\\\nI1203 13:48:06.938315 6697 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938336 6697 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 13:48:06.938394 6697 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:48:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.813207 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.823522 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.836145 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:08Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.890220 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.890262 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.890273 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.890289 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.890300 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.976031 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.976148 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:08 crc kubenswrapper[4677]: E1203 13:48:08.976273 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.976336 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:08 crc kubenswrapper[4677]: E1203 13:48:08.976440 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:08 crc kubenswrapper[4677]: E1203 13:48:08.976477 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.993227 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.993279 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.993291 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.993305 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:08 crc kubenswrapper[4677]: I1203 13:48:08.993315 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:08Z","lastTransitionTime":"2025-12-03T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.095735 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.095780 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.095791 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.095805 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.095816 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.199077 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.199458 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.199714 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.199933 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.200192 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.303057 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.303110 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.303123 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.303143 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.303157 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.405303 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.405349 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.405367 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.405388 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.405402 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.507710 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.507781 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.507799 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.507823 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.507845 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.609935 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.609990 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.609999 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.610012 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.610024 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.712161 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.712192 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.712202 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.712214 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.712223 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.814816 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.814899 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.814925 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.814990 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.815019 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.917670 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.917749 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.917773 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.917802 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.917823 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:09Z","lastTransitionTime":"2025-12-03T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.975556 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:09 crc kubenswrapper[4677]: E1203 13:48:09.975769 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:09 crc kubenswrapper[4677]: I1203 13:48:09.992622 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a027c7aa-3b4c-447d-bf1b-70ba613d2d42\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f5f20f4d82df343a33abd3c75925824d24cecf691f0f7e785f3ca480e390d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1793451a6eb08865434705a0d6ddea74bb2bbafd92c74730a8de43bff6b7c68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fcc15539cd9399027479c6a3eba342a978b9d2bbe3d4ed3b07ee656480e342f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:09Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.005932 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xjvnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76e85245-8632-41e9-b4bd-987b708709a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee4d33be169612253040d8d5edafa3f2237b160b9778b38436256565636d36e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kv6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xjvnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.020401 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.020450 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.020466 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.020487 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.020505 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.027573 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88e2b75b-db00-4a56-8e0b-0dc9abb880d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 13:47:03.176647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 13:47:03.177577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-245170094/tls.crt::/tmp/serving-cert-245170094/tls.key\\\\\\\"\\\\nI1203 13:47:08.678689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 13:47:08.684801 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 13:47:08.684838 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 13:47:08.684861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 13:47:08.684870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 13:47:08.691454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 13:47:08.691477 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 13:47:08.691499 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 13:47:08.691527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 13:47:08.691539 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 13:47:08.691550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 13:47:08.691559 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 13:47:08.695040 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.043138 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.058243 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3532ebbcc8a4c883a9c5f94c35bf85c9376cf965a25360a8e2b6ab0f202cdf80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.073717 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b1323e42de0da7bce574355fde371af29594fcbc96d43566b3667a322c766cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.089146 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j7h62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcgcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j7h62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.107761 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f3f2e8-5876-4a54-896b-0985bc0df323\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55c800949358da2923f6aa0298858b275f483be88b806ae633ac31df53df621c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07fb3635227f990ca4919031ae5cc207fb130dfa32e8a8e0d248ee12952119d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58527e906f44722bd0ee7d56ae69b4adc3ff1240175479909b55f55665f217a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10208d429f4e2770a0ec36319958a1399e094da0bfde0540f19336a33a4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b0a955dad2d790c11b24d2fbbf5d94e1887d7671e6a516294c8f4c81daf26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f23694da0f071dacd09798b99adb8617ea0cc5121a92c95108a75561b11669f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc7d49356ba4718fc6eb94b0af0d1a68dcdf397640491521f6705e14afe8ba4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b200fa2b6e54274b4fdb2cac45e84a0d956edcfb7d0239669042d20c1137bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.123668 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"190fcd64-a72b-492c-a572-74a7bf265478\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://190691d05ab29233c3a62564cd09bcfbfe7e738a9fb87bd4e2f39c83b51f328a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f54818b597438c08a372ac009489eb2e3e500a6549f103bced004248eac44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e23746222dad684d25bd8068e2296c798e7b746f23e2bc6865d3c2094c39a824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acd17011d74035f24fbaefb48516346bc05924c3c060b06201ebb271e79fd618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.123807 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.123859 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.123874 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.123895 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.123910 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.140603 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://476620c579b320ad2de481aa12787591697aa93578d5600c55606a19eb6311f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.153739 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.175040 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaf2e40d-0316-4380-961f-8039e0674f2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:48:07Z\\\",\\\"message\\\":\\\"ultus/network-metrics-daemon-j7h62 openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql openshift-ovn-kubernetes/ovnkube-node-pm7pz]\\\\nI1203 13:48:06.938254 6697 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1203 13:48:06.938274 6697 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938287 6697 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938298 6697 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz in node crc\\\\nI1203 13:48:06.938306 6697 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pm7pz after 0 failed attempt(s)\\\\nI1203 13:48:06.938315 6697 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pm7pz\\\\nI1203 13:48:06.938336 6697 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 13:48:06.938394 6697 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:48:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lxth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pm7pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.186127 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"545df6df-eecf-4d8b-8981-0b57b96b227a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6b1564728698cf898fe304d5d83d2071f624d749ac85f0a07206af3cb4610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803c8895795b1cc1cc357b693f44b1fe0738b613b2b781dd711aa8902729e570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.197431 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.207255 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1665842-4668-4fed-a04f-cf4aa0043ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f850f373fe6a29a34ed81954df53923051f6d47286b9af4f25d788a796386cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf2m2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2bx7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.223205 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7nch7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511a6f65-7dac-4f37-a15e-3a24339f80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T13:47:58Z\\\",\\\"message\\\":\\\"2025-12-03T13:47:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54\\\\n2025-12-03T13:47:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_34dad215-129b-4b09-961e-054f6ff79a54 to /host/opt/cni/bin/\\\\n2025-12-03T13:47:13Z [verbose] multus-daemon started\\\\n2025-12-03T13:47:13Z [verbose] Readiness Indicator file check\\\\n2025-12-03T13:47:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhct7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7nch7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.227232 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.227290 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.227303 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.227322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.227334 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.241789 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sxc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0755a37-6619-4dd8-b598-d2bf05c3a3ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef4a33570c2eb2a84cdbca0c7ceddfafef42448e7f52a7f39e691819284896ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1171a6e098aa77d6fd0cbf9783f44484597aa7b9e58e6b490be9792b4c112977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d6759650d8877ed671f5a427ebdb40a6687c4db163b137dec6d3ad2080c88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fa8c7719a757b3f443b3fe8eedf36a52865b28dc93fb310bd7c3ad200b6aea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83347c33199137171cd3ec76c0fa4de1f54cf2946317ed94df10e17dd3d70c00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5676f45528a2e11fa929b5a5b619863354cc462e71f484141ba9153948296cb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61cc5ab46ede8438bc0c1bb3cd0d622eff50adde1f13ceebdeb8e9d086ab5532\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T13:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T13:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sxc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.252496 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdbzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cfbbe23-6cec-425e-a192-10c5f7f36317\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2135241ae86d198605bbde5cb6456b8551c447d7223140731580aa0ce8e2e574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn7qj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdbzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.263531 4677 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97274f10-85e1-4b88-b061-36427e54a9d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T13:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51879ecaa5376e46e3f685330f2e22de5dfe2329259e6f28605241787fbe29d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544835f8da2876bb21e8650f5ba3a031032fae7fc02bd55f99c045f335a29eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T13:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nj2m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T13:47:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mjkql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T13:48:10Z is after 2025-08-24T17:21:41Z" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.329810 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.329854 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.329868 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.329884 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.329895 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.432885 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.432932 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.432942 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.432976 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.432988 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.536213 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.536277 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.536292 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.536319 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.536376 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.639723 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.639767 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.639780 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.639797 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.639810 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.743130 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.743200 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.743217 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.743244 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.743264 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.846641 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.846693 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.846707 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.846727 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.846741 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.949359 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.949413 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.949425 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.949444 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.949456 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:10Z","lastTransitionTime":"2025-12-03T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.975742 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.975824 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:10 crc kubenswrapper[4677]: I1203 13:48:10.975738 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:10 crc kubenswrapper[4677]: E1203 13:48:10.975872 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:10 crc kubenswrapper[4677]: E1203 13:48:10.976043 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:10 crc kubenswrapper[4677]: E1203 13:48:10.976074 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.053118 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.053201 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.053235 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.053262 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.053276 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.156369 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.156448 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.156471 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.156500 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.156521 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.259755 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.259821 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.259835 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.259857 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.259872 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.363047 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.363524 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.363723 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.363907 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.364061 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.467759 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.468332 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.468482 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.468700 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.468940 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.572555 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.572805 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.572906 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.573102 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.573249 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.676874 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.676937 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.677048 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.677073 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.677089 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.779817 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.779913 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.779930 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.780071 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.780100 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.883184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.883244 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.883312 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.883338 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.883355 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.975476 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:11 crc kubenswrapper[4677]: E1203 13:48:11.975708 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.985931 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.986021 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.986044 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.986071 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:11 crc kubenswrapper[4677]: I1203 13:48:11.986093 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:11Z","lastTransitionTime":"2025-12-03T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.088802 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.088866 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.088887 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.088913 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.088928 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.190939 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.191039 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.191052 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.191070 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.191083 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.294444 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.294506 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.294520 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.294538 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.294550 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.397228 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.397538 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.397652 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.397748 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.397845 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.501187 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.501267 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.501292 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.501322 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.501345 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.603468 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.603578 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.603626 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.603647 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.603663 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.706880 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.706915 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.706924 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.706978 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.706989 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.809588 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.809632 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.809640 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.809654 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.809663 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.829334 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.829533 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.829755 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:16.829727452 +0000 UTC m=+147.576059917 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.829798 4677 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.830048 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:49:16.830032532 +0000 UTC m=+147.576364997 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.912773 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.912810 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.912823 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.912841 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.912854 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:12Z","lastTransitionTime":"2025-12-03T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.930312 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.930379 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.930441 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930564 4677 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930622 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 13:49:16.930603156 +0000 UTC m=+147.676935631 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930629 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930675 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930678 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930700 4677 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930722 4677 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930747 4677 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930779 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 13:49:16.93075487 +0000 UTC m=+147.677087365 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.930824 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 13:49:16.930796792 +0000 UTC m=+147.677129297 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.975873 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.976039 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.976194 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.976332 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:12 crc kubenswrapper[4677]: I1203 13:48:12.976422 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:12 crc kubenswrapper[4677]: E1203 13:48:12.976586 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.016185 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.016243 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.016261 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.016284 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.016302 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.118922 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.119039 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.119072 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.119101 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.119123 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.222718 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.222852 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.222893 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.222925 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.222996 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.325452 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.325500 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.325515 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.325535 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.325568 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.427979 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.428036 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.428053 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.428075 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.428091 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.530812 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.530870 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.530886 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.530910 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.530929 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.633199 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.633250 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.633264 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.633279 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.633291 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.736286 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.736343 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.736357 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.736378 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.736392 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.839036 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.839114 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.839123 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.839137 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.839147 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.941608 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.942010 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.942175 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.942301 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.942424 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:13Z","lastTransitionTime":"2025-12-03T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:13 crc kubenswrapper[4677]: I1203 13:48:13.975478 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:13 crc kubenswrapper[4677]: E1203 13:48:13.975666 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.044783 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.045017 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.045028 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.045042 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.045052 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:14Z","lastTransitionTime":"2025-12-03T13:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.148097 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.148149 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.148161 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.148184 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.148200 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:14Z","lastTransitionTime":"2025-12-03T13:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.251343 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.251428 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.251457 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.251487 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.251505 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:14Z","lastTransitionTime":"2025-12-03T13:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.355025 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.355106 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.355131 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.355162 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.355184 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:14Z","lastTransitionTime":"2025-12-03T13:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.458115 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.458170 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.458179 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.458193 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.458204 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:14Z","lastTransitionTime":"2025-12-03T13:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.478201 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.478242 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.478250 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.478264 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.478273 4677 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T13:48:14Z","lastTransitionTime":"2025-12-03T13:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.528402 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff"] Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.528807 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.530655 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.530730 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.531980 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.532374 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.546361 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.546407 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.546429 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.546445 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.546487 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.549938 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=61.549917798 podStartE2EDuration="1m1.549917798s" podCreationTimestamp="2025-12-03 13:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.549868806 +0000 UTC m=+85.296201271" watchObservedRunningTime="2025-12-03 13:48:14.549917798 +0000 UTC m=+85.296250243" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.564430 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xjvnv" podStartSLOduration=64.564416743 podStartE2EDuration="1m4.564416743s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.564250787 +0000 UTC m=+85.310583252" watchObservedRunningTime="2025-12-03 13:48:14.564416743 +0000 UTC m=+85.310749198" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.601375 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=66.601358696 podStartE2EDuration="1m6.601358696s" podCreationTimestamp="2025-12-03 13:47:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.589699343 +0000 UTC m=+85.336031818" watchObservedRunningTime="2025-12-03 13:48:14.601358696 +0000 UTC m=+85.347691151" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.647763 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.647823 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.647849 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.647875 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.647913 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.647923 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.647920 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.648710 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.661844 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.665493 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afe82e2-a6c3-478b-afd5-f8e5ce209a1f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dvsff\" (UID: \"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.681608 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=62.681586068 podStartE2EDuration="1m2.681586068s" podCreationTimestamp="2025-12-03 13:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.680996309 +0000 UTC m=+85.427328784" watchObservedRunningTime="2025-12-03 13:48:14.681586068 +0000 UTC m=+85.427918523" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.706142 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=30.706127374 podStartE2EDuration="30.706127374s" podCreationTimestamp="2025-12-03 13:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.694375168 +0000 UTC m=+85.440707653" watchObservedRunningTime="2025-12-03 13:48:14.706127374 +0000 UTC m=+85.452459849" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.750912 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-sxc65" podStartSLOduration=64.750894669 podStartE2EDuration="1m4.750894669s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.750174456 +0000 UTC m=+85.496506941" watchObservedRunningTime="2025-12-03 13:48:14.750894669 +0000 UTC m=+85.497227124" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.771499 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-bdbzd" podStartSLOduration=64.77148188 podStartE2EDuration="1m4.77148188s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.770554879 +0000 UTC m=+85.516887334" watchObservedRunningTime="2025-12-03 13:48:14.77148188 +0000 UTC m=+85.517814345" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.794509 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mjkql" podStartSLOduration=64.794489977 podStartE2EDuration="1m4.794489977s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.781750648 +0000 UTC m=+85.528083103" watchObservedRunningTime="2025-12-03 13:48:14.794489977 +0000 UTC m=+85.540822432" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.804885 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=22.80486966 podStartE2EDuration="22.80486966s" podCreationTimestamp="2025-12-03 13:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.804437956 +0000 UTC m=+85.550770411" watchObservedRunningTime="2025-12-03 13:48:14.80486966 +0000 UTC m=+85.551202115" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.827102 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podStartSLOduration=64.827084182 podStartE2EDuration="1m4.827084182s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.826325277 +0000 UTC m=+85.572657742" watchObservedRunningTime="2025-12-03 13:48:14.827084182 +0000 UTC m=+85.573416637" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.840080 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-7nch7" podStartSLOduration=64.840063847 podStartE2EDuration="1m4.840063847s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:14.839039765 +0000 UTC m=+85.585372240" watchObservedRunningTime="2025-12-03 13:48:14.840063847 +0000 UTC m=+85.586396302" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.844131 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" Dec 03 13:48:14 crc kubenswrapper[4677]: W1203 13:48:14.863170 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7afe82e2_a6c3_478b_afd5_f8e5ce209a1f.slice/crio-c2be3524bae8022068dda85da43b4d8efc1a968a10244f64e34550d3535ee47d WatchSource:0}: Error finding container c2be3524bae8022068dda85da43b4d8efc1a968a10244f64e34550d3535ee47d: Status 404 returned error can't find the container with id c2be3524bae8022068dda85da43b4d8efc1a968a10244f64e34550d3535ee47d Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.975711 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.975749 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:14 crc kubenswrapper[4677]: I1203 13:48:14.975752 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:14 crc kubenswrapper[4677]: E1203 13:48:14.975831 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:14 crc kubenswrapper[4677]: E1203 13:48:14.975877 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:14 crc kubenswrapper[4677]: E1203 13:48:14.975992 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:15 crc kubenswrapper[4677]: I1203 13:48:15.635107 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" event={"ID":"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f","Type":"ContainerStarted","Data":"e063b4bf3008676a47462f0e42c2c5c1b48c4ca9d87988a563e5da6769b2dff0"} Dec 03 13:48:15 crc kubenswrapper[4677]: I1203 13:48:15.635170 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" event={"ID":"7afe82e2-a6c3-478b-afd5-f8e5ce209a1f","Type":"ContainerStarted","Data":"c2be3524bae8022068dda85da43b4d8efc1a968a10244f64e34550d3535ee47d"} Dec 03 13:48:15 crc kubenswrapper[4677]: I1203 13:48:15.653379 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvsff" podStartSLOduration=65.653360866 podStartE2EDuration="1m5.653360866s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:15.652613761 +0000 UTC m=+86.398946226" watchObservedRunningTime="2025-12-03 13:48:15.653360866 +0000 UTC m=+86.399693321" Dec 03 13:48:15 crc kubenswrapper[4677]: I1203 13:48:15.975404 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:15 crc kubenswrapper[4677]: E1203 13:48:15.975581 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:16 crc kubenswrapper[4677]: I1203 13:48:16.975375 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:16 crc kubenswrapper[4677]: I1203 13:48:16.975431 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:16 crc kubenswrapper[4677]: I1203 13:48:16.975410 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:16 crc kubenswrapper[4677]: E1203 13:48:16.975541 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:16 crc kubenswrapper[4677]: E1203 13:48:16.975754 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:16 crc kubenswrapper[4677]: E1203 13:48:16.975858 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:17 crc kubenswrapper[4677]: I1203 13:48:17.976233 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:17 crc kubenswrapper[4677]: E1203 13:48:17.976701 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:18 crc kubenswrapper[4677]: I1203 13:48:18.975729 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:18 crc kubenswrapper[4677]: I1203 13:48:18.975790 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:18 crc kubenswrapper[4677]: E1203 13:48:18.975874 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:18 crc kubenswrapper[4677]: E1203 13:48:18.976061 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:18 crc kubenswrapper[4677]: I1203 13:48:18.976390 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:18 crc kubenswrapper[4677]: E1203 13:48:18.976617 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:19 crc kubenswrapper[4677]: I1203 13:48:19.975512 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:19 crc kubenswrapper[4677]: E1203 13:48:19.978686 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:20 crc kubenswrapper[4677]: I1203 13:48:20.974942 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:20 crc kubenswrapper[4677]: E1203 13:48:20.975323 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:20 crc kubenswrapper[4677]: I1203 13:48:20.975054 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:20 crc kubenswrapper[4677]: E1203 13:48:20.975545 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:20 crc kubenswrapper[4677]: I1203 13:48:20.975042 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:20 crc kubenswrapper[4677]: E1203 13:48:20.976142 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:21 crc kubenswrapper[4677]: I1203 13:48:21.975613 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:21 crc kubenswrapper[4677]: E1203 13:48:21.975812 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:22 crc kubenswrapper[4677]: I1203 13:48:22.975506 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:22 crc kubenswrapper[4677]: I1203 13:48:22.975506 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:22 crc kubenswrapper[4677]: E1203 13:48:22.975690 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:22 crc kubenswrapper[4677]: I1203 13:48:22.975517 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:22 crc kubenswrapper[4677]: E1203 13:48:22.975791 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:22 crc kubenswrapper[4677]: E1203 13:48:22.976379 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:22 crc kubenswrapper[4677]: I1203 13:48:22.976469 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 13:48:22 crc kubenswrapper[4677]: E1203 13:48:22.976689 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:48:23 crc kubenswrapper[4677]: I1203 13:48:23.975654 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:23 crc kubenswrapper[4677]: E1203 13:48:23.975794 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:24 crc kubenswrapper[4677]: I1203 13:48:24.976071 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:24 crc kubenswrapper[4677]: I1203 13:48:24.976106 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:24 crc kubenswrapper[4677]: I1203 13:48:24.976106 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:24 crc kubenswrapper[4677]: E1203 13:48:24.976419 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:24 crc kubenswrapper[4677]: E1203 13:48:24.976439 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:24 crc kubenswrapper[4677]: E1203 13:48:24.976474 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:25 crc kubenswrapper[4677]: I1203 13:48:25.975231 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:25 crc kubenswrapper[4677]: E1203 13:48:25.975438 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:26 crc kubenswrapper[4677]: I1203 13:48:26.975274 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:26 crc kubenswrapper[4677]: I1203 13:48:26.975364 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:26 crc kubenswrapper[4677]: I1203 13:48:26.975304 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:26 crc kubenswrapper[4677]: E1203 13:48:26.975543 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:26 crc kubenswrapper[4677]: E1203 13:48:26.975682 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:26 crc kubenswrapper[4677]: E1203 13:48:26.976255 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:27 crc kubenswrapper[4677]: I1203 13:48:27.976022 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:27 crc kubenswrapper[4677]: E1203 13:48:27.976162 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:28 crc kubenswrapper[4677]: I1203 13:48:28.975564 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:28 crc kubenswrapper[4677]: I1203 13:48:28.975728 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:28 crc kubenswrapper[4677]: E1203 13:48:28.975762 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:28 crc kubenswrapper[4677]: I1203 13:48:28.975579 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:28 crc kubenswrapper[4677]: E1203 13:48:28.976011 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:28 crc kubenswrapper[4677]: E1203 13:48:28.976170 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:29 crc kubenswrapper[4677]: I1203 13:48:29.005461 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:29 crc kubenswrapper[4677]: E1203 13:48:29.005687 4677 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:48:29 crc kubenswrapper[4677]: E1203 13:48:29.005802 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs podName:d26430e4-fe9f-4b2f-ae90-a91fd8fccf79 nodeName:}" failed. No retries permitted until 2025-12-03 13:49:33.00577833 +0000 UTC m=+163.752110805 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs") pod "network-metrics-daemon-j7h62" (UID: "d26430e4-fe9f-4b2f-ae90-a91fd8fccf79") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 13:48:29 crc kubenswrapper[4677]: I1203 13:48:29.975402 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:29 crc kubenswrapper[4677]: E1203 13:48:29.976585 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:30 crc kubenswrapper[4677]: I1203 13:48:30.975397 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:30 crc kubenswrapper[4677]: I1203 13:48:30.975486 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:30 crc kubenswrapper[4677]: I1203 13:48:30.975510 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:30 crc kubenswrapper[4677]: E1203 13:48:30.975592 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:30 crc kubenswrapper[4677]: E1203 13:48:30.975682 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:30 crc kubenswrapper[4677]: E1203 13:48:30.975793 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:31 crc kubenswrapper[4677]: I1203 13:48:31.975849 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:31 crc kubenswrapper[4677]: E1203 13:48:31.976143 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:32 crc kubenswrapper[4677]: I1203 13:48:32.975423 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:32 crc kubenswrapper[4677]: I1203 13:48:32.975480 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:32 crc kubenswrapper[4677]: I1203 13:48:32.975526 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:32 crc kubenswrapper[4677]: E1203 13:48:32.975707 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:32 crc kubenswrapper[4677]: E1203 13:48:32.975857 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:32 crc kubenswrapper[4677]: E1203 13:48:32.976089 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:33 crc kubenswrapper[4677]: I1203 13:48:33.975439 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:33 crc kubenswrapper[4677]: E1203 13:48:33.975673 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:34 crc kubenswrapper[4677]: I1203 13:48:34.975654 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:34 crc kubenswrapper[4677]: I1203 13:48:34.975686 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:34 crc kubenswrapper[4677]: I1203 13:48:34.975750 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:34 crc kubenswrapper[4677]: E1203 13:48:34.975808 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:34 crc kubenswrapper[4677]: E1203 13:48:34.975884 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:34 crc kubenswrapper[4677]: E1203 13:48:34.976074 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:35 crc kubenswrapper[4677]: I1203 13:48:35.976101 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:35 crc kubenswrapper[4677]: E1203 13:48:35.976434 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:36 crc kubenswrapper[4677]: I1203 13:48:36.975423 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:36 crc kubenswrapper[4677]: I1203 13:48:36.975532 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:36 crc kubenswrapper[4677]: I1203 13:48:36.975450 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:36 crc kubenswrapper[4677]: E1203 13:48:36.975605 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:36 crc kubenswrapper[4677]: E1203 13:48:36.975754 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:36 crc kubenswrapper[4677]: E1203 13:48:36.975928 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:36 crc kubenswrapper[4677]: I1203 13:48:36.976717 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 13:48:36 crc kubenswrapper[4677]: E1203 13:48:36.976895 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pm7pz_openshift-ovn-kubernetes(eaf2e40d-0316-4380-961f-8039e0674f2c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" Dec 03 13:48:37 crc kubenswrapper[4677]: I1203 13:48:37.976242 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:37 crc kubenswrapper[4677]: E1203 13:48:37.976454 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:38 crc kubenswrapper[4677]: I1203 13:48:38.975010 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:38 crc kubenswrapper[4677]: I1203 13:48:38.975084 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:38 crc kubenswrapper[4677]: E1203 13:48:38.975118 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:38 crc kubenswrapper[4677]: E1203 13:48:38.975255 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:38 crc kubenswrapper[4677]: I1203 13:48:38.976139 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:38 crc kubenswrapper[4677]: E1203 13:48:38.976222 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:39 crc kubenswrapper[4677]: I1203 13:48:39.976120 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:39 crc kubenswrapper[4677]: E1203 13:48:39.978012 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:40 crc kubenswrapper[4677]: I1203 13:48:40.975806 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:40 crc kubenswrapper[4677]: I1203 13:48:40.975922 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:40 crc kubenswrapper[4677]: E1203 13:48:40.975984 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:40 crc kubenswrapper[4677]: I1203 13:48:40.975807 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:40 crc kubenswrapper[4677]: E1203 13:48:40.976107 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:40 crc kubenswrapper[4677]: E1203 13:48:40.976290 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:41 crc kubenswrapper[4677]: I1203 13:48:41.975667 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:41 crc kubenswrapper[4677]: E1203 13:48:41.975836 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:42 crc kubenswrapper[4677]: I1203 13:48:42.975885 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:42 crc kubenswrapper[4677]: I1203 13:48:42.975913 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:42 crc kubenswrapper[4677]: I1203 13:48:42.976034 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:42 crc kubenswrapper[4677]: E1203 13:48:42.976232 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:42 crc kubenswrapper[4677]: E1203 13:48:42.976411 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:42 crc kubenswrapper[4677]: E1203 13:48:42.976575 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:43 crc kubenswrapper[4677]: I1203 13:48:43.975558 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:43 crc kubenswrapper[4677]: E1203 13:48:43.975688 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:44 crc kubenswrapper[4677]: I1203 13:48:44.975221 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:44 crc kubenswrapper[4677]: I1203 13:48:44.975273 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:44 crc kubenswrapper[4677]: E1203 13:48:44.975350 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:44 crc kubenswrapper[4677]: I1203 13:48:44.975360 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:44 crc kubenswrapper[4677]: E1203 13:48:44.975513 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:44 crc kubenswrapper[4677]: E1203 13:48:44.975605 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:45 crc kubenswrapper[4677]: I1203 13:48:45.751425 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/1.log" Dec 03 13:48:45 crc kubenswrapper[4677]: I1203 13:48:45.752823 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/0.log" Dec 03 13:48:45 crc kubenswrapper[4677]: I1203 13:48:45.752888 4677 generic.go:334] "Generic (PLEG): container finished" podID="511a6f65-7dac-4f37-a15e-3a24339f80f4" containerID="6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c" exitCode=1 Dec 03 13:48:45 crc kubenswrapper[4677]: I1203 13:48:45.752935 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerDied","Data":"6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c"} Dec 03 13:48:45 crc kubenswrapper[4677]: I1203 13:48:45.753012 4677 scope.go:117] "RemoveContainer" containerID="0762e14256d9d6271226c8bccf574d54a855072fc56a2d4da96aa8901fc54d19" Dec 03 13:48:45 crc kubenswrapper[4677]: I1203 13:48:45.753583 4677 scope.go:117] "RemoveContainer" containerID="6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c" Dec 03 13:48:45 crc kubenswrapper[4677]: E1203 13:48:45.753846 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-7nch7_openshift-multus(511a6f65-7dac-4f37-a15e-3a24339f80f4)\"" pod="openshift-multus/multus-7nch7" podUID="511a6f65-7dac-4f37-a15e-3a24339f80f4" Dec 03 13:48:45 crc kubenswrapper[4677]: I1203 13:48:45.975638 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:45 crc kubenswrapper[4677]: E1203 13:48:45.975829 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:46 crc kubenswrapper[4677]: I1203 13:48:46.758479 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/1.log" Dec 03 13:48:46 crc kubenswrapper[4677]: I1203 13:48:46.975623 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:46 crc kubenswrapper[4677]: I1203 13:48:46.975684 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:46 crc kubenswrapper[4677]: E1203 13:48:46.975775 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:46 crc kubenswrapper[4677]: I1203 13:48:46.975918 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:46 crc kubenswrapper[4677]: E1203 13:48:46.976198 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:46 crc kubenswrapper[4677]: E1203 13:48:46.976327 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:47 crc kubenswrapper[4677]: I1203 13:48:47.975274 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:47 crc kubenswrapper[4677]: E1203 13:48:47.975424 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:48 crc kubenswrapper[4677]: I1203 13:48:48.975907 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:48 crc kubenswrapper[4677]: E1203 13:48:48.976039 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:48 crc kubenswrapper[4677]: I1203 13:48:48.976199 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:48 crc kubenswrapper[4677]: E1203 13:48:48.976355 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:48 crc kubenswrapper[4677]: I1203 13:48:48.977562 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:48 crc kubenswrapper[4677]: E1203 13:48:48.977881 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:49 crc kubenswrapper[4677]: E1203 13:48:49.928498 4677 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 13:48:49 crc kubenswrapper[4677]: I1203 13:48:49.975839 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:49 crc kubenswrapper[4677]: E1203 13:48:49.976817 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:50 crc kubenswrapper[4677]: E1203 13:48:50.075155 4677 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:48:50 crc kubenswrapper[4677]: I1203 13:48:50.975078 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:50 crc kubenswrapper[4677]: I1203 13:48:50.975178 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:50 crc kubenswrapper[4677]: I1203 13:48:50.975178 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:50 crc kubenswrapper[4677]: E1203 13:48:50.975740 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:50 crc kubenswrapper[4677]: E1203 13:48:50.975889 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:50 crc kubenswrapper[4677]: I1203 13:48:50.975923 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 13:48:50 crc kubenswrapper[4677]: E1203 13:48:50.976097 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:51 crc kubenswrapper[4677]: I1203 13:48:51.780784 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/3.log" Dec 03 13:48:51 crc kubenswrapper[4677]: I1203 13:48:51.784511 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerStarted","Data":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} Dec 03 13:48:51 crc kubenswrapper[4677]: I1203 13:48:51.784987 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:48:51 crc kubenswrapper[4677]: I1203 13:48:51.819464 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podStartSLOduration=101.819444078 podStartE2EDuration="1m41.819444078s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:48:51.817481385 +0000 UTC m=+122.563813850" watchObservedRunningTime="2025-12-03 13:48:51.819444078 +0000 UTC m=+122.565776543" Dec 03 13:48:51 crc kubenswrapper[4677]: I1203 13:48:51.976429 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:51 crc kubenswrapper[4677]: E1203 13:48:51.976969 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:51 crc kubenswrapper[4677]: I1203 13:48:51.982863 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-j7h62"] Dec 03 13:48:51 crc kubenswrapper[4677]: I1203 13:48:51.982943 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:51 crc kubenswrapper[4677]: E1203 13:48:51.983040 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:52 crc kubenswrapper[4677]: I1203 13:48:52.975936 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:52 crc kubenswrapper[4677]: I1203 13:48:52.976029 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:52 crc kubenswrapper[4677]: E1203 13:48:52.976196 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:52 crc kubenswrapper[4677]: E1203 13:48:52.976296 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:53 crc kubenswrapper[4677]: I1203 13:48:53.975995 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:53 crc kubenswrapper[4677]: E1203 13:48:53.976398 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:53 crc kubenswrapper[4677]: I1203 13:48:53.975999 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:53 crc kubenswrapper[4677]: E1203 13:48:53.976982 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:54 crc kubenswrapper[4677]: I1203 13:48:54.975004 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:54 crc kubenswrapper[4677]: I1203 13:48:54.975061 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:54 crc kubenswrapper[4677]: E1203 13:48:54.975383 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:54 crc kubenswrapper[4677]: E1203 13:48:54.975519 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:55 crc kubenswrapper[4677]: E1203 13:48:55.077224 4677 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:48:55 crc kubenswrapper[4677]: I1203 13:48:55.975822 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:55 crc kubenswrapper[4677]: E1203 13:48:55.976044 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:55 crc kubenswrapper[4677]: I1203 13:48:55.975822 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:55 crc kubenswrapper[4677]: E1203 13:48:55.976359 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:56 crc kubenswrapper[4677]: I1203 13:48:56.975668 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:56 crc kubenswrapper[4677]: E1203 13:48:56.975791 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:56 crc kubenswrapper[4677]: I1203 13:48:56.975670 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:56 crc kubenswrapper[4677]: E1203 13:48:56.975884 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:57 crc kubenswrapper[4677]: I1203 13:48:57.975120 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:57 crc kubenswrapper[4677]: I1203 13:48:57.975153 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:57 crc kubenswrapper[4677]: E1203 13:48:57.975276 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:48:57 crc kubenswrapper[4677]: E1203 13:48:57.975440 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:58 crc kubenswrapper[4677]: I1203 13:48:58.975231 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:48:58 crc kubenswrapper[4677]: E1203 13:48:58.975405 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:48:58 crc kubenswrapper[4677]: I1203 13:48:58.975251 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:48:58 crc kubenswrapper[4677]: E1203 13:48:58.975569 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:48:59 crc kubenswrapper[4677]: I1203 13:48:59.976048 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:48:59 crc kubenswrapper[4677]: E1203 13:48:59.978024 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:48:59 crc kubenswrapper[4677]: I1203 13:48:59.978048 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:48:59 crc kubenswrapper[4677]: E1203 13:48:59.978177 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:49:00 crc kubenswrapper[4677]: E1203 13:49:00.078640 4677 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 13:49:00 crc kubenswrapper[4677]: I1203 13:49:00.975163 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:00 crc kubenswrapper[4677]: I1203 13:49:00.975167 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:00 crc kubenswrapper[4677]: E1203 13:49:00.975554 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:49:00 crc kubenswrapper[4677]: E1203 13:49:00.975821 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:49:00 crc kubenswrapper[4677]: I1203 13:49:00.976138 4677 scope.go:117] "RemoveContainer" containerID="6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c" Dec 03 13:49:01 crc kubenswrapper[4677]: I1203 13:49:01.817515 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/1.log" Dec 03 13:49:01 crc kubenswrapper[4677]: I1203 13:49:01.817567 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerStarted","Data":"0b27c752b379608243e6b1a21d41b90e4bf7e320e5eab9403c29cd5a04921959"} Dec 03 13:49:01 crc kubenswrapper[4677]: I1203 13:49:01.976253 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:49:01 crc kubenswrapper[4677]: I1203 13:49:01.976269 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:49:01 crc kubenswrapper[4677]: E1203 13:49:01.976454 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:49:01 crc kubenswrapper[4677]: E1203 13:49:01.976529 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:49:02 crc kubenswrapper[4677]: I1203 13:49:02.976030 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:02 crc kubenswrapper[4677]: E1203 13:49:02.976174 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:49:02 crc kubenswrapper[4677]: I1203 13:49:02.976045 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:02 crc kubenswrapper[4677]: E1203 13:49:02.976353 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:49:03 crc kubenswrapper[4677]: I1203 13:49:03.975778 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:49:03 crc kubenswrapper[4677]: I1203 13:49:03.975868 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:49:03 crc kubenswrapper[4677]: E1203 13:49:03.975913 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 13:49:03 crc kubenswrapper[4677]: E1203 13:49:03.976070 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j7h62" podUID="d26430e4-fe9f-4b2f-ae90-a91fd8fccf79" Dec 03 13:49:04 crc kubenswrapper[4677]: I1203 13:49:04.975541 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:04 crc kubenswrapper[4677]: E1203 13:49:04.975680 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 13:49:04 crc kubenswrapper[4677]: I1203 13:49:04.975540 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:04 crc kubenswrapper[4677]: E1203 13:49:04.975766 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 13:49:05 crc kubenswrapper[4677]: I1203 13:49:05.344255 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 13:49:05 crc kubenswrapper[4677]: I1203 13:49:05.975372 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:49:05 crc kubenswrapper[4677]: I1203 13:49:05.975425 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:49:05 crc kubenswrapper[4677]: I1203 13:49:05.979131 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 13:49:05 crc kubenswrapper[4677]: I1203 13:49:05.979418 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 13:49:05 crc kubenswrapper[4677]: I1203 13:49:05.979511 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 13:49:05 crc kubenswrapper[4677]: I1203 13:49:05.980260 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 13:49:06 crc kubenswrapper[4677]: I1203 13:49:06.975904 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:06 crc kubenswrapper[4677]: I1203 13:49:06.975904 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:06 crc kubenswrapper[4677]: I1203 13:49:06.978938 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 13:49:06 crc kubenswrapper[4677]: I1203 13:49:06.979366 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.090371 4677 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.156761 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gd9tn"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.157693 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.157742 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ppzxd"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.158226 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.158944 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d4p7j"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.159453 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.160981 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.161520 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.162042 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.162657 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.163153 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntqz8"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.163809 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.164224 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-s84pq"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.164563 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.165346 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.166083 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.171153 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-55xx9"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.171635 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.171965 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-66lfj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.185389 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.185553 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.185656 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.185770 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.190841 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.190887 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.191084 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.192599 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.191348 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.191454 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.193858 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.194285 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.195124 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.195232 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xxds8"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.195667 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.196232 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.196389 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.196250 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.197229 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.213354 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.213588 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.213738 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8qnws"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.213826 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.213905 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.213998 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214130 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214587 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93a0fb86-4f19-4023-b777-c58381af5047-images\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214684 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-audit\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214711 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22d27e8-6223-4370-9c02-ef73ec251631-serving-cert\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214730 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-etcd-serving-ca\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214746 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-serving-cert\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214770 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-image-import-ca\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214786 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdq78\" (UniqueName: \"kubernetes.io/projected/93a0fb86-4f19-4023-b777-c58381af5047-kube-api-access-kdq78\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214804 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214819 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-etcd-client\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214834 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7e33975-80e2-410a-8819-3d66e41ccd21-audit-dir\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214887 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214918 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215009 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215278 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.214926 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n2rl\" (UniqueName: \"kubernetes.io/projected/b22d27e8-6223-4370-9c02-ef73ec251631-kube-api-access-5n2rl\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215566 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f7e33975-80e2-410a-8819-3d66e41ccd21-node-pullsecrets\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215598 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215611 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-config\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215638 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7sd4\" (UniqueName: \"kubernetes.io/projected/f7e33975-80e2-410a-8819-3d66e41ccd21-kube-api-access-r7sd4\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215663 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-client-ca\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215688 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a0fb86-4f19-4023-b777-c58381af5047-config\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215712 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/93a0fb86-4f19-4023-b777-c58381af5047-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215737 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-encryption-config\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.215757 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-config\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.216295 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.216415 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.216622 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.216726 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.216814 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.216912 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217047 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217058 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217096 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217151 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217170 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217266 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217280 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217300 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217382 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217463 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217473 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217587 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217648 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217663 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217756 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217852 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217925 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218028 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218123 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218408 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218517 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218580 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218695 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218764 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218806 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218878 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.219025 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.219243 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.219501 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.219624 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217053 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.219747 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.217856 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.219854 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.219875 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.218831 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.220017 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.220025 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.220241 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.221402 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hct2b"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.221890 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.222593 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.222993 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223267 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223403 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223492 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223511 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223586 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223600 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223655 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223719 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223770 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223824 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223866 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.223332 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.224282 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.228779 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.229789 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.230203 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.231047 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.231440 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.231043 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-t9z7p"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.231991 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.232313 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.232521 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-t9z7p" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.232702 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.232818 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.233986 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.234500 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.234560 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.235020 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.235022 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.238818 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.239374 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.239736 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.239773 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.252728 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.254289 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.255586 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-hmrqp"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.255807 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.256286 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.258482 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.258646 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.258730 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.259408 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.259525 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.258565 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.274344 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.275370 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.274532 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.274689 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.275973 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.276119 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.277294 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.278681 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lpf4d"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.279103 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.280342 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.281165 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.282017 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.282373 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-74nhw"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.282555 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.282785 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.283736 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-xmbtm"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.283995 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.284209 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lqjhf"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.284252 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.288300 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.290336 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.290509 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.291659 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.291730 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.293425 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-p4cw4"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.293427 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.293491 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.295737 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.296069 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.296314 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.296422 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.297121 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.297255 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.297763 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.297866 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.298778 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.298892 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.299259 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.299409 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ppzxd"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.299521 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.299577 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntqz8"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.300183 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.301870 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gd9tn"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.308161 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.308205 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-66lfj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.308219 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.313891 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.316856 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318342 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-config\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318394 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7sd4\" (UniqueName: \"kubernetes.io/projected/f7e33975-80e2-410a-8819-3d66e41ccd21-kube-api-access-r7sd4\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318430 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-client-ca\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318464 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318495 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jszbh\" (UniqueName: \"kubernetes.io/projected/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-kube-api-access-jszbh\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318523 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a0fb86-4f19-4023-b777-c58381af5047-config\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318550 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/93a0fb86-4f19-4023-b777-c58381af5047-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318581 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-audit-policies\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318646 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-encryption-config\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318673 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-config\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318704 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318733 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318761 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93a0fb86-4f19-4023-b777-c58381af5047-images\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318820 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-audit-dir\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318849 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-audit\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318885 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318929 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22d27e8-6223-4370-9c02-ef73ec251631-serving-cert\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.318975 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-etcd-serving-ca\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319017 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-serving-cert\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319050 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-image-import-ca\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319082 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319111 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdq78\" (UniqueName: \"kubernetes.io/projected/93a0fb86-4f19-4023-b777-c58381af5047-kube-api-access-kdq78\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319142 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-serving-cert\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319171 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-encryption-config\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319202 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319237 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-etcd-client\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319262 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7e33975-80e2-410a-8819-3d66e41ccd21-audit-dir\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319309 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319334 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb2ff\" (UniqueName: \"kubernetes.io/projected/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-kube-api-access-tb2ff\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319368 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n2rl\" (UniqueName: \"kubernetes.io/projected/b22d27e8-6223-4370-9c02-ef73ec251631-kube-api-access-5n2rl\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319396 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-etcd-client\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319439 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f7e33975-80e2-410a-8819-3d66e41ccd21-node-pullsecrets\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.319589 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f7e33975-80e2-410a-8819-3d66e41ccd21-node-pullsecrets\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.321440 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-config\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.322063 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-client-ca\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.323136 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a0fb86-4f19-4023-b777-c58381af5047-config\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.323851 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-etcd-serving-ca\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.324385 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.324443 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-r8sn4"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.324858 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.327054 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-image-import-ca\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.328485 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8qnws"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.328631 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.330555 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xxds8"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.330565 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93a0fb86-4f19-4023-b777-c58381af5047-images\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.331114 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7e33975-80e2-410a-8819-3d66e41ccd21-audit-dir\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.331135 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-audit\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.332312 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7e33975-80e2-410a-8819-3d66e41ccd21-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.332834 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hct2b"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.334851 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22d27e8-6223-4370-9c02-ef73ec251631-serving-cert\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.337268 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.337634 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-etcd-client\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.339569 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.342866 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.343215 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s84pq"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.343260 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lpf4d"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.344387 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lqjhf"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.345638 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/93a0fb86-4f19-4023-b777-c58381af5047-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.347078 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.347431 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-encryption-config\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.347690 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7e33975-80e2-410a-8819-3d66e41ccd21-serving-cert\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.355153 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-config\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.356493 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.359439 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d4p7j"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.359513 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.360535 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-xmbtm"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.362963 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.367104 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.369841 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.373576 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.375373 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.377261 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-55xx9"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.380262 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.380408 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-t9z7p"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.381609 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.384276 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-p4cw4"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.385767 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-74nhw"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.387924 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.390292 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.391723 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.393323 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.395202 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.396701 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ftj4m"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.398030 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.398253 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bq6q6"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.398906 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.399745 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bq6q6"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.400746 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.400974 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ftj4m"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420475 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420518 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jszbh\" (UniqueName: \"kubernetes.io/projected/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-kube-api-access-jszbh\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420543 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-audit-policies\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420576 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420600 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420637 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-audit-dir\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420666 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420717 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420746 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-serving-cert\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420763 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-encryption-config\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420786 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb2ff\" (UniqueName: \"kubernetes.io/projected/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-kube-api-access-tb2ff\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.420818 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-etcd-client\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.423523 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-audit-dir\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.424396 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-audit-policies\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.424786 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.425808 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.430365 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-serving-cert\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.430748 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.434022 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.434822 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.436173 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-etcd-client\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.439245 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-encryption-config\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.441811 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.443073 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-ltrs7"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.443932 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.448335 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ltrs7"] Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.460613 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.479614 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.500736 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.520703 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.540593 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.565832 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.579604 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.599640 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.620423 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.641303 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.660528 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.681419 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.700239 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.720579 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.739738 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.759444 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.781047 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.800913 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.820142 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.840150 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.860934 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.880750 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.900463 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.921002 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 13:49:15 crc kubenswrapper[4677]: I1203 13:49:15.941053 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.001142 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.021045 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.040327 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.060301 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.080574 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.100472 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.120402 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.140034 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.160910 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.180750 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.200713 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.220400 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.240728 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.260660 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.281092 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.298191 4677 request.go:700] Waited for 1.015152598s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.299649 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.320182 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.340874 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.360775 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.380315 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.399768 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.420658 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.440326 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.460525 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.480652 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.500816 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.520483 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.539666 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.564998 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.580036 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.600279 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.620249 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.642295 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.660793 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.680531 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.700304 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.720845 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.739797 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.760183 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.780424 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.800724 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.820627 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.835275 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:16 crc kubenswrapper[4677]: E1203 13:49:16.835387 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:51:18.835365119 +0000 UTC m=+269.581697574 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.835418 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.836099 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.840050 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.860316 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.881005 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.900063 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.919759 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.937345 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.937438 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.937503 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.939747 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.941017 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.941032 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.941676 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.976608 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7sd4\" (UniqueName: \"kubernetes.io/projected/f7e33975-80e2-410a-8819-3d66e41ccd21-kube-api-access-r7sd4\") pod \"apiserver-76f77b778f-gd9tn\" (UID: \"f7e33975-80e2-410a-8819-3d66e41ccd21\") " pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.985278 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.995330 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdq78\" (UniqueName: \"kubernetes.io/projected/93a0fb86-4f19-4023-b777-c58381af5047-kube-api-access-kdq78\") pod \"machine-api-operator-5694c8668f-d4p7j\" (UID: \"93a0fb86-4f19-4023-b777-c58381af5047\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:16 crc kubenswrapper[4677]: I1203 13:49:16.999628 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.020252 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.040115 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.054577 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.076923 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n2rl\" (UniqueName: \"kubernetes.io/projected/b22d27e8-6223-4370-9c02-ef73ec251631-kube-api-access-5n2rl\") pod \"controller-manager-879f6c89f-ppzxd\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.079643 4677 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.100000 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.101831 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.123154 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.140287 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.161111 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.182748 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.193145 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.198928 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.225893 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gd9tn"] Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.227022 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.240871 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jszbh\" (UniqueName: \"kubernetes.io/projected/3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885-kube-api-access-jszbh\") pod \"apiserver-7bbb656c7d-z92ld\" (UID: \"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.257496 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d4p7j"] Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.260797 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb2ff\" (UniqueName: \"kubernetes.io/projected/80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a-kube-api-access-tb2ff\") pod \"cluster-image-registry-operator-dc59b4c8b-hqdkl\" (UID: \"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.260893 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 13:49:17 crc kubenswrapper[4677]: W1203 13:49:17.273438 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93a0fb86_4f19_4023_b777_c58381af5047.slice/crio-9890f3f73eaafd68af6bd5254980900c666c7a16b7e04351e5047ff25f6a1832 WatchSource:0}: Error finding container 9890f3f73eaafd68af6bd5254980900c666c7a16b7e04351e5047ff25f6a1832: Status 404 returned error can't find the container with id 9890f3f73eaafd68af6bd5254980900c666c7a16b7e04351e5047ff25f6a1832 Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.279981 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.302118 4677 request.go:700] Waited for 1.857657621s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.304446 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.320047 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.320335 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.341838 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.341894 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.341922 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0963f4ad-95a1-4602-ab12-3073f3db0581-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342003 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-bound-sa-token\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342032 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d97cr\" (UniqueName: \"kubernetes.io/projected/d91c7be8-82e6-4da1-b483-67f45cac28ef-kube-api-access-d97cr\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342057 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2c20cb8-d990-4756-ad31-56245b583279-audit-dir\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342272 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.342312 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:17.842299067 +0000 UTC m=+148.588631532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342499 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342733 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0963f4ad-95a1-4602-ab12-3073f3db0581-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342802 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d91c7be8-82e6-4da1-b483-67f45cac28ef-images\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.342825 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.343533 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91c7be8-82e6-4da1-b483-67f45cac28ef-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.343710 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-config\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.343784 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-audit-policies\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.343810 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.343918 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsj5r\" (UniqueName: \"kubernetes.io/projected/c2c20cb8-d990-4756-ad31-56245b583279-kube-api-access-hsj5r\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344182 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-certificates\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344213 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344280 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344302 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-service-ca\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344359 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344381 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-client\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344401 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-ca\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344427 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344447 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344469 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtvh4\" (UniqueName: \"kubernetes.io/projected/60a15680-2160-4cea-a7e0-0a3a22685b3f-kube-api-access-qtvh4\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344854 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60a15680-2160-4cea-a7e0-0a3a22685b3f-serving-cert\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.344909 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d91c7be8-82e6-4da1-b483-67f45cac28ef-proxy-tls\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.345228 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwzpd\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-kube-api-access-kwzpd\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.346071 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-tls\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.346440 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-trusted-ca\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.346492 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.382143 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.447342 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.447408 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.447430 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c84b5\" (UniqueName: \"kubernetes.io/projected/a265d87b-7caf-4b02-8afe-388cef7069b2-kube-api-access-c84b5\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.447445 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngfpz\" (UniqueName: \"kubernetes.io/projected/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-kube-api-access-ngfpz\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449100 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449135 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqnb4\" (UniqueName: \"kubernetes.io/projected/515d4b6a-61f4-4929-b3be-d152d8ced855-kube-api-access-vqnb4\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449153 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0963f4ad-95a1-4602-ab12-3073f3db0581-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.449190 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:17.949151402 +0000 UTC m=+148.695483857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449230 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20bba33b-274c-4099-ae84-4fc906ffc07b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449275 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbm7c\" (UniqueName: \"kubernetes.io/projected/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-kube-api-access-pbm7c\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449291 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2djz\" (UniqueName: \"kubernetes.io/projected/4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9-kube-api-access-r2djz\") pod \"ingress-canary-ltrs7\" (UID: \"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9\") " pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449343 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d97cr\" (UniqueName: \"kubernetes.io/projected/d91c7be8-82e6-4da1-b483-67f45cac28ef-kube-api-access-d97cr\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449360 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk5pv\" (UniqueName: \"kubernetes.io/projected/e1149c46-0323-453e-a042-4a6e8155364d-kube-api-access-nk5pv\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449377 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgbhr\" (UniqueName: \"kubernetes.io/projected/9c0a9578-40a5-4345-8a4b-3e809f978d48-kube-api-access-pgbhr\") pod \"downloads-7954f5f757-t9z7p\" (UID: \"9c0a9578-40a5-4345-8a4b-3e809f978d48\") " pod="openshift-console/downloads-7954f5f757-t9z7p" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449408 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0b9868b-191a-4a71-b538-e67f0dbaa54f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449425 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01c58bf7-d620-4770-bfc1-b63c29cf32f9-trusted-ca\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449442 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a265d87b-7caf-4b02-8afe-388cef7069b2-trusted-ca\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449462 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12ffe862-7453-46fe-8aae-95065549f876-metrics-tls\") pod \"dns-operator-744455d44c-lpf4d\" (UID: \"12ffe862-7453-46fe-8aae-95065549f876\") " pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449515 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8231653d-6197-4f1b-85bf-8f5c260d47fe-service-ca-bundle\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449531 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-metrics-certs\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449545 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/515d4b6a-61f4-4929-b3be-d152d8ced855-signing-cabundle\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449581 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449598 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd64d\" (UniqueName: \"kubernetes.io/projected/5e893d3c-7f15-44a6-b881-f658c0e06171-kube-api-access-xd64d\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449614 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93eeaace-9fca-4e02-909a-3db0b7ab56e4-proxy-tls\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449648 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-plugins-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449678 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0963f4ad-95a1-4602-ab12-3073f3db0581-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449694 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449725 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d91c7be8-82e6-4da1-b483-67f45cac28ef-images\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449741 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-config\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449764 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-config\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449795 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pstf8\" (UniqueName: \"kubernetes.io/projected/2214172a-784b-4626-9e7d-e0a7a0bbfeb7-kube-api-access-pstf8\") pod \"package-server-manager-789f6589d5-gbcsj\" (UID: \"2214172a-784b-4626-9e7d-e0a7a0bbfeb7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449815 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-config\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449832 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84259910-9b22-4957-ad88-7d00052ee57c-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449849 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwn2k\" (UniqueName: \"kubernetes.io/projected/da2727c5-131f-460a-82a9-60e440d73aef-kube-api-access-bwn2k\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449890 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-srv-cert\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449908 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/da040c14-6dff-4426-8258-6bd0e3e14cb6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2c9gq\" (UID: \"da040c14-6dff-4426-8258-6bd0e3e14cb6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449926 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8fr5\" (UniqueName: \"kubernetes.io/projected/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-kube-api-access-d8fr5\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449966 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-service-ca-bundle\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.449983 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-profile-collector-cert\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450001 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450132 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da2727c5-131f-460a-82a9-60e440d73aef-serving-cert\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450225 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450243 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-client-ca\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450258 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f5t2\" (UniqueName: \"kubernetes.io/projected/01c58bf7-d620-4770-bfc1-b63c29cf32f9-kube-api-access-8f5t2\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450289 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-service-ca\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450306 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqzt7\" (UniqueName: \"kubernetes.io/projected/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-kube-api-access-gqzt7\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450322 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4f4d19d6-6b48-4c81-85d9-36a3c076d498-certs\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450337 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-csi-data-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450377 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450393 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-socket-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450409 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450425 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450456 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjn59\" (UniqueName: \"kubernetes.io/projected/20bba33b-274c-4099-ae84-4fc906ffc07b-kube-api-access-rjn59\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450475 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtvh4\" (UniqueName: \"kubernetes.io/projected/60a15680-2160-4cea-a7e0-0a3a22685b3f-kube-api-access-qtvh4\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450491 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c5c6f36-9c21-487e-85f3-633d81390939-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bhtkj\" (UID: \"2c5c6f36-9c21-487e-85f3-633d81390939\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450508 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wmgk\" (UniqueName: \"kubernetes.io/projected/4f4d19d6-6b48-4c81-85d9-36a3c076d498-kube-api-access-2wmgk\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450540 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lztcv\" (UniqueName: \"kubernetes.io/projected/84259910-9b22-4957-ad88-7d00052ee57c-kube-api-access-lztcv\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450556 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60a15680-2160-4cea-a7e0-0a3a22685b3f-serving-cert\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450587 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2214172a-784b-4626-9e7d-e0a7a0bbfeb7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gbcsj\" (UID: \"2214172a-784b-4626-9e7d-e0a7a0bbfeb7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450620 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwzpd\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-kube-api-access-kwzpd\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450658 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-tls\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450690 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-stats-auth\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450706 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-service-ca\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450720 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qgn7\" (UniqueName: \"kubernetes.io/projected/b693f37f-a924-4f57-a4a5-c9ba03815229-kube-api-access-4qgn7\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450762 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450779 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-serving-cert\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450796 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-config\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450836 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ff07b87-8826-40b5-a881-9cdabc473090-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450865 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a265d87b-7caf-4b02-8afe-388cef7069b2-metrics-tls\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.450880 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/515d4b6a-61f4-4929-b3be-d152d8ced855-signing-key\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451013 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4f4d19d6-6b48-4c81-85d9-36a3c076d498-node-bootstrap-token\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451042 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9a04339-599f-4f20-a2bf-9d12addcd11f-config-volume\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451089 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp5c6\" (UniqueName: \"kubernetes.io/projected/f9a04339-599f-4f20-a2bf-9d12addcd11f-kube-api-access-kp5c6\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451105 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqw5k\" (UniqueName: \"kubernetes.io/projected/58b0e34a-a6be-4a51-9ef4-4dd2234ec15f-kube-api-access-bqw5k\") pod \"multus-admission-controller-857f4d67dd-p4cw4\" (UID: \"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451122 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/735074c0-8634-44aa-9cfe-cdd1a07633d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451193 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-registration-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451734 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451757 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z47xp\" (UniqueName: \"kubernetes.io/projected/12ffe862-7453-46fe-8aae-95065549f876-kube-api-access-z47xp\") pod \"dns-operator-744455d44c-lpf4d\" (UID: \"12ffe862-7453-46fe-8aae-95065549f876\") " pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451795 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451815 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d8ce3e58-f4e8-444f-ba23-172619e34134-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451887 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.451912 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-bound-sa-token\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.452025 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e1149c46-0323-453e-a042-4a6e8155364d-webhook-cert\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.452046 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2c20cb8-d990-4756-ad31-56245b583279-audit-dir\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.452075 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97wxc\" (UniqueName: \"kubernetes.io/projected/8231653d-6197-4f1b-85bf-8f5c260d47fe-kube-api-access-97wxc\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.452113 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/58b0e34a-a6be-4a51-9ef4-4dd2234ec15f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-p4cw4\" (UID: \"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.452129 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-mountpoint-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.452177 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.452356 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0963f4ad-95a1-4602-ab12-3073f3db0581-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453001 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-service-ca\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453235 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d91c7be8-82e6-4da1-b483-67f45cac28ef-images\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453795 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-config\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453818 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e893d3c-7f15-44a6-b881-f658c0e06171-config-volume\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453837 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91c7be8-82e6-4da1-b483-67f45cac28ef-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453856 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-trusted-ca-bundle\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453868 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2c20cb8-d990-4756-ad31-56245b583279-audit-dir\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453877 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-audit-policies\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453895 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453912 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0b9868b-191a-4a71-b538-e67f0dbaa54f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453931 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmvks\" (UniqueName: \"kubernetes.io/projected/93eeaace-9fca-4e02-909a-3db0b7ab56e4-kube-api-access-fmvks\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.453990 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2727c5-131f-460a-82a9-60e440d73aef-config\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454030 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e1149c46-0323-453e-a042-4a6e8155364d-apiservice-cert\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454047 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c58bf7-d620-4770-bfc1-b63c29cf32f9-config\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454067 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsj5r\" (UniqueName: \"kubernetes.io/projected/c2c20cb8-d990-4756-ad31-56245b583279-kube-api-access-hsj5r\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454086 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8jrz\" (UniqueName: \"kubernetes.io/projected/d8ce3e58-f4e8-444f-ba23-172619e34134-kube-api-access-m8jrz\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454301 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4sgb\" (UniqueName: \"kubernetes.io/projected/9dc01e2f-388f-41a2-a139-5d251ad6cda8-kube-api-access-z4sgb\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454321 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r47j4\" (UniqueName: \"kubernetes.io/projected/735074c0-8634-44aa-9cfe-cdd1a07633d4-kube-api-access-r47j4\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454339 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-certificates\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454523 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454624 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-config\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.454646 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:17.954629748 +0000 UTC m=+148.700962203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454693 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20bba33b-274c-4099-ae84-4fc906ffc07b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454717 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93eeaace-9fca-4e02-909a-3db0b7ab56e4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454752 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhpl2\" (UniqueName: \"kubernetes.io/projected/894546c9-1858-4460-8a89-02cc798d7c4c-kube-api-access-mhpl2\") pod \"migrator-59844c95c7-l5kb9\" (UID: \"894546c9-1858-4460-8a89-02cc798d7c4c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454769 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e1149c46-0323-453e-a042-4a6e8155364d-tmpfs\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454788 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-client\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454806 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-machine-approver-tls\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454821 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-config\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454837 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9a04339-599f-4f20-a2bf-9d12addcd11f-metrics-tls\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454853 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454871 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-ca\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454887 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-auth-proxy-config\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454903 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454920 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a265d87b-7caf-4b02-8afe-388cef7069b2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454936 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8ce3e58-f4e8-444f-ba23-172619e34134-serving-cert\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.454965 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.455339 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456279 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60a15680-2160-4cea-a7e0-0a3a22685b3f-serving-cert\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456545 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-certificates\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456595 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456630 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d91c7be8-82e6-4da1-b483-67f45cac28ef-proxy-tls\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456648 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c58bf7-d620-4770-bfc1-b63c29cf32f9-serving-cert\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456665 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ff07b87-8826-40b5-a881-9cdabc473090-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456687 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-default-certificate\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456703 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-oauth-config\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456718 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ff07b87-8826-40b5-a881-9cdabc473090-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456735 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9-cert\") pod \"ingress-canary-ltrs7\" (UID: \"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9\") " pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.456789 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457091 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457493 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457563 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-oauth-serving-cert\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457589 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9glj\" (UniqueName: \"kubernetes.io/projected/da040c14-6dff-4426-8258-6bd0e3e14cb6-kube-api-access-f9glj\") pod \"control-plane-machine-set-operator-78cbb6b69f-2c9gq\" (UID: \"da040c14-6dff-4426-8258-6bd0e3e14cb6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457617 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-trusted-ca\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457669 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nbvm\" (UniqueName: \"kubernetes.io/projected/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-kube-api-access-8nbvm\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457690 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b9868b-191a-4a71-b538-e67f0dbaa54f-config\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457711 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcqpg\" (UniqueName: \"kubernetes.io/projected/2c5c6f36-9c21-487e-85f3-633d81390939-kube-api-access-hcqpg\") pod \"cluster-samples-operator-665b6dd947-bhtkj\" (UID: \"2c5c6f36-9c21-487e-85f3-633d81390939\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457735 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b693f37f-a924-4f57-a4a5-c9ba03815229-serving-cert\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457754 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e893d3c-7f15-44a6-b881-f658c0e06171-secret-volume\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457781 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/735074c0-8634-44aa-9cfe-cdd1a07633d4-srv-cert\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.457903 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.458005 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.458098 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t79qg\" (UniqueName: \"kubernetes.io/projected/dfeedbc7-41ac-4853-9254-7b8876109abf-kube-api-access-t79qg\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.458688 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91c7be8-82e6-4da1-b483-67f45cac28ef-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.458720 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-ca\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.458936 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60a15680-2160-4cea-a7e0-0a3a22685b3f-etcd-client\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.459325 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-trusted-ca\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.459788 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-audit-policies\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.459902 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.461610 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.462143 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.462896 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0963f4ad-95a1-4602-ab12-3073f3db0581-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.463530 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-tls\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.463579 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d91c7be8-82e6-4da1-b483-67f45cac28ef-proxy-tls\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.464053 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.466468 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.466870 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.488722 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.494759 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtvh4\" (UniqueName: \"kubernetes.io/projected/60a15680-2160-4cea-a7e0-0a3a22685b3f-kube-api-access-qtvh4\") pod \"etcd-operator-b45778765-55xx9\" (UID: \"60a15680-2160-4cea-a7e0-0a3a22685b3f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.499723 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ppzxd"] Dec 03 13:49:17 crc kubenswrapper[4677]: W1203 13:49:17.512513 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb22d27e8_6223_4370_9c02_ef73ec251631.slice/crio-efad9c818ccff235f6b00795f3679034b3617c6741a5b5b6be28a8c69d2f2445 WatchSource:0}: Error finding container efad9c818ccff235f6b00795f3679034b3617c6741a5b5b6be28a8c69d2f2445: Status 404 returned error can't find the container with id efad9c818ccff235f6b00795f3679034b3617c6741a5b5b6be28a8c69d2f2445 Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.515206 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d97cr\" (UniqueName: \"kubernetes.io/projected/d91c7be8-82e6-4da1-b483-67f45cac28ef-kube-api-access-d97cr\") pod \"machine-config-operator-74547568cd-qmq88\" (UID: \"d91c7be8-82e6-4da1-b483-67f45cac28ef\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.537976 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwzpd\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-kube-api-access-kwzpd\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.558081 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-bound-sa-token\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571319 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.571464 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.071444562 +0000 UTC m=+148.817777027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571544 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571569 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z47xp\" (UniqueName: \"kubernetes.io/projected/12ffe862-7453-46fe-8aae-95065549f876-kube-api-access-z47xp\") pod \"dns-operator-744455d44c-lpf4d\" (UID: \"12ffe862-7453-46fe-8aae-95065549f876\") " pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571586 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571605 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d8ce3e58-f4e8-444f-ba23-172619e34134-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571627 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571645 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e1149c46-0323-453e-a042-4a6e8155364d-webhook-cert\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571663 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97wxc\" (UniqueName: \"kubernetes.io/projected/8231653d-6197-4f1b-85bf-8f5c260d47fe-kube-api-access-97wxc\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571684 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/58b0e34a-a6be-4a51-9ef4-4dd2234ec15f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-p4cw4\" (UID: \"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571705 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-mountpoint-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571724 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-config\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571744 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e893d3c-7f15-44a6-b881-f658c0e06171-config-volume\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571762 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-trusted-ca-bundle\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571783 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2727c5-131f-460a-82a9-60e440d73aef-config\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571804 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0b9868b-191a-4a71-b538-e67f0dbaa54f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571827 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmvks\" (UniqueName: \"kubernetes.io/projected/93eeaace-9fca-4e02-909a-3db0b7ab56e4-kube-api-access-fmvks\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571848 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e1149c46-0323-453e-a042-4a6e8155364d-apiservice-cert\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571869 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c58bf7-d620-4770-bfc1-b63c29cf32f9-config\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571897 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8jrz\" (UniqueName: \"kubernetes.io/projected/d8ce3e58-f4e8-444f-ba23-172619e34134-kube-api-access-m8jrz\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571917 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4sgb\" (UniqueName: \"kubernetes.io/projected/9dc01e2f-388f-41a2-a139-5d251ad6cda8-kube-api-access-z4sgb\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.571938 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r47j4\" (UniqueName: \"kubernetes.io/projected/735074c0-8634-44aa-9cfe-cdd1a07633d4-kube-api-access-r47j4\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572038 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20bba33b-274c-4099-ae84-4fc906ffc07b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572096 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93eeaace-9fca-4e02-909a-3db0b7ab56e4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572123 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhpl2\" (UniqueName: \"kubernetes.io/projected/894546c9-1858-4460-8a89-02cc798d7c4c-kube-api-access-mhpl2\") pod \"migrator-59844c95c7-l5kb9\" (UID: \"894546c9-1858-4460-8a89-02cc798d7c4c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572192 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e1149c46-0323-453e-a042-4a6e8155364d-tmpfs\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572214 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-machine-approver-tls\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572233 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-config\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572255 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9a04339-599f-4f20-a2bf-9d12addcd11f-metrics-tls\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572366 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572389 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-auth-proxy-config\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572409 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572432 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a265d87b-7caf-4b02-8afe-388cef7069b2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572451 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8ce3e58-f4e8-444f-ba23-172619e34134-serving-cert\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572473 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572495 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c58bf7-d620-4770-bfc1-b63c29cf32f9-serving-cert\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572517 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ff07b87-8826-40b5-a881-9cdabc473090-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572538 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-default-certificate\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572557 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-oauth-config\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572578 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ff07b87-8826-40b5-a881-9cdabc473090-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572596 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9-cert\") pod \"ingress-canary-ltrs7\" (UID: \"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9\") " pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572613 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-oauth-serving-cert\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572634 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9glj\" (UniqueName: \"kubernetes.io/projected/da040c14-6dff-4426-8258-6bd0e3e14cb6-kube-api-access-f9glj\") pod \"control-plane-machine-set-operator-78cbb6b69f-2c9gq\" (UID: \"da040c14-6dff-4426-8258-6bd0e3e14cb6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572656 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nbvm\" (UniqueName: \"kubernetes.io/projected/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-kube-api-access-8nbvm\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572673 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b9868b-191a-4a71-b538-e67f0dbaa54f-config\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572694 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcqpg\" (UniqueName: \"kubernetes.io/projected/2c5c6f36-9c21-487e-85f3-633d81390939-kube-api-access-hcqpg\") pod \"cluster-samples-operator-665b6dd947-bhtkj\" (UID: \"2c5c6f36-9c21-487e-85f3-633d81390939\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572714 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b693f37f-a924-4f57-a4a5-c9ba03815229-serving-cert\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572733 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e893d3c-7f15-44a6-b881-f658c0e06171-secret-volume\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572751 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/735074c0-8634-44aa-9cfe-cdd1a07633d4-srv-cert\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572770 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572788 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t79qg\" (UniqueName: \"kubernetes.io/projected/dfeedbc7-41ac-4853-9254-7b8876109abf-kube-api-access-t79qg\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572806 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572826 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c84b5\" (UniqueName: \"kubernetes.io/projected/a265d87b-7caf-4b02-8afe-388cef7069b2-kube-api-access-c84b5\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572845 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngfpz\" (UniqueName: \"kubernetes.io/projected/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-kube-api-access-ngfpz\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572867 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqnb4\" (UniqueName: \"kubernetes.io/projected/515d4b6a-61f4-4929-b3be-d152d8ced855-kube-api-access-vqnb4\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572891 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20bba33b-274c-4099-ae84-4fc906ffc07b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572913 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbm7c\" (UniqueName: \"kubernetes.io/projected/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-kube-api-access-pbm7c\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572933 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2djz\" (UniqueName: \"kubernetes.io/projected/4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9-kube-api-access-r2djz\") pod \"ingress-canary-ltrs7\" (UID: \"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9\") " pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.572977 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk5pv\" (UniqueName: \"kubernetes.io/projected/e1149c46-0323-453e-a042-4a6e8155364d-kube-api-access-nk5pv\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.573001 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgbhr\" (UniqueName: \"kubernetes.io/projected/9c0a9578-40a5-4345-8a4b-3e809f978d48-kube-api-access-pgbhr\") pod \"downloads-7954f5f757-t9z7p\" (UID: \"9c0a9578-40a5-4345-8a4b-3e809f978d48\") " pod="openshift-console/downloads-7954f5f757-t9z7p" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.573020 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0b9868b-191a-4a71-b538-e67f0dbaa54f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.573040 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01c58bf7-d620-4770-bfc1-b63c29cf32f9-trusted-ca\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.573060 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a265d87b-7caf-4b02-8afe-388cef7069b2-trusted-ca\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.573079 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12ffe862-7453-46fe-8aae-95065549f876-metrics-tls\") pod \"dns-operator-744455d44c-lpf4d\" (UID: \"12ffe862-7453-46fe-8aae-95065549f876\") " pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.574520 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8231653d-6197-4f1b-85bf-8f5c260d47fe-service-ca-bundle\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.574522 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-config\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.574600 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-mountpoint-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.575337 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da2727c5-131f-460a-82a9-60e440d73aef-config\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.573098 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8231653d-6197-4f1b-85bf-8f5c260d47fe-service-ca-bundle\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.576225 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.076204865 +0000 UTC m=+148.822537320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576285 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-metrics-certs\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576304 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/515d4b6a-61f4-4929-b3be-d152d8ced855-signing-cabundle\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576349 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd64d\" (UniqueName: \"kubernetes.io/projected/5e893d3c-7f15-44a6-b881-f658c0e06171-kube-api-access-xd64d\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576367 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93eeaace-9fca-4e02-909a-3db0b7ab56e4-proxy-tls\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576415 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-plugins-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576446 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-config\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576464 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-config\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576504 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pstf8\" (UniqueName: \"kubernetes.io/projected/2214172a-784b-4626-9e7d-e0a7a0bbfeb7-kube-api-access-pstf8\") pod \"package-server-manager-789f6589d5-gbcsj\" (UID: \"2214172a-784b-4626-9e7d-e0a7a0bbfeb7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576523 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84259910-9b22-4957-ad88-7d00052ee57c-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576539 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwn2k\" (UniqueName: \"kubernetes.io/projected/da2727c5-131f-460a-82a9-60e440d73aef-kube-api-access-bwn2k\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576577 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-srv-cert\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576595 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/da040c14-6dff-4426-8258-6bd0e3e14cb6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2c9gq\" (UID: \"da040c14-6dff-4426-8258-6bd0e3e14cb6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576618 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8fr5\" (UniqueName: \"kubernetes.io/projected/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-kube-api-access-d8fr5\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576655 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-service-ca-bundle\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576675 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-profile-collector-cert\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.576692 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579699 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da2727c5-131f-460a-82a9-60e440d73aef-serving-cert\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579726 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-client-ca\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579745 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f5t2\" (UniqueName: \"kubernetes.io/projected/01c58bf7-d620-4770-bfc1-b63c29cf32f9-kube-api-access-8f5t2\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579764 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqzt7\" (UniqueName: \"kubernetes.io/projected/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-kube-api-access-gqzt7\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579781 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4f4d19d6-6b48-4c81-85d9-36a3c076d498-certs\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579801 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-csi-data-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579818 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579836 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-socket-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579927 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjn59\" (UniqueName: \"kubernetes.io/projected/20bba33b-274c-4099-ae84-4fc906ffc07b-kube-api-access-rjn59\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579964 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c5c6f36-9c21-487e-85f3-633d81390939-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bhtkj\" (UID: \"2c5c6f36-9c21-487e-85f3-633d81390939\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579980 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wmgk\" (UniqueName: \"kubernetes.io/projected/4f4d19d6-6b48-4c81-85d9-36a3c076d498-kube-api-access-2wmgk\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.579999 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lztcv\" (UniqueName: \"kubernetes.io/projected/84259910-9b22-4957-ad88-7d00052ee57c-kube-api-access-lztcv\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580016 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2214172a-784b-4626-9e7d-e0a7a0bbfeb7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gbcsj\" (UID: \"2214172a-784b-4626-9e7d-e0a7a0bbfeb7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580039 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-stats-auth\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580056 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-service-ca\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580073 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qgn7\" (UniqueName: \"kubernetes.io/projected/b693f37f-a924-4f57-a4a5-c9ba03815229-kube-api-access-4qgn7\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580091 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-serving-cert\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580109 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-config\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580124 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ff07b87-8826-40b5-a881-9cdabc473090-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580143 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a265d87b-7caf-4b02-8afe-388cef7069b2-metrics-tls\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580158 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/515d4b6a-61f4-4929-b3be-d152d8ced855-signing-key\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580173 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4f4d19d6-6b48-4c81-85d9-36a3c076d498-node-bootstrap-token\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580190 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9a04339-599f-4f20-a2bf-9d12addcd11f-config-volume\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580207 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp5c6\" (UniqueName: \"kubernetes.io/projected/f9a04339-599f-4f20-a2bf-9d12addcd11f-kube-api-access-kp5c6\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580225 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqw5k\" (UniqueName: \"kubernetes.io/projected/58b0e34a-a6be-4a51-9ef4-4dd2234ec15f-kube-api-access-bqw5k\") pod \"multus-admission-controller-857f4d67dd-p4cw4\" (UID: \"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580243 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/735074c0-8634-44aa-9cfe-cdd1a07633d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.580258 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-registration-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.581023 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-trusted-ca-bundle\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.581353 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-registration-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.589454 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/515d4b6a-61f4-4929-b3be-d152d8ced855-signing-cabundle\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.591682 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-service-ca-bundle\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.591828 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-oauth-config\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.592404 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-metrics-certs\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.593163 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/da040c14-6dff-4426-8258-6bd0e3e14cb6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2c9gq\" (UID: \"da040c14-6dff-4426-8258-6bd0e3e14cb6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.593205 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84259910-9b22-4957-ad88-7d00052ee57c-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.593829 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e1149c46-0323-453e-a042-4a6e8155364d-webhook-cert\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.594263 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/58b0e34a-a6be-4a51-9ef4-4dd2234ec15f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-p4cw4\" (UID: \"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.594590 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.596092 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsj5r\" (UniqueName: \"kubernetes.io/projected/c2c20cb8-d990-4756-ad31-56245b583279-kube-api-access-hsj5r\") pod \"oauth-openshift-558db77b4-ntqz8\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.597810 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-srv-cert\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.598799 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-config\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.598903 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-plugins-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.600389 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-config\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.600677 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2214172a-784b-4626-9e7d-e0a7a0bbfeb7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gbcsj\" (UID: \"2214172a-784b-4626-9e7d-e0a7a0bbfeb7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.578881 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld"] Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.600803 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93eeaace-9fca-4e02-909a-3db0b7ab56e4-proxy-tls\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.578932 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e893d3c-7f15-44a6-b881-f658c0e06171-config-volume\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.602980 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-socket-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.603244 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ff07b87-8826-40b5-a881-9cdabc473090-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.604068 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20bba33b-274c-4099-ae84-4fc906ffc07b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.604148 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9-cert\") pod \"ingress-canary-ltrs7\" (UID: \"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9\") " pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.604270 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.605031 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93eeaace-9fca-4e02-909a-3db0b7ab56e4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.605473 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-service-ca\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.608702 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-config\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.609200 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ff07b87-8826-40b5-a881-9cdabc473090-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.610507 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9a04339-599f-4f20-a2bf-9d12addcd11f-config-volume\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.610682 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e1149c46-0323-453e-a042-4a6e8155364d-tmpfs\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.610916 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/515d4b6a-61f4-4929-b3be-d152d8ced855-signing-key\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.611620 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-stats-auth\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.613047 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c58bf7-d620-4770-bfc1-b63c29cf32f9-config\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.613688 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a265d87b-7caf-4b02-8afe-388cef7069b2-metrics-tls\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.614242 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4f4d19d6-6b48-4c81-85d9-36a3c076d498-node-bootstrap-token\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.614486 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-profile-collector-cert\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.614707 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dfeedbc7-41ac-4853-9254-7b8876109abf-csi-data-dir\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.616546 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-serving-cert\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.617725 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4f4d19d6-6b48-4c81-85d9-36a3c076d498-certs\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.617936 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.618162 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.618349 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-config\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.619193 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b693f37f-a924-4f57-a4a5-c9ba03815229-serving-cert\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.619284 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-oauth-serving-cert\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.619428 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-machine-approver-tls\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.620987 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20bba33b-274c-4099-ae84-4fc906ffc07b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.621074 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-client-ca\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.621533 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da2727c5-131f-460a-82a9-60e440d73aef-serving-cert\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.619777 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.621692 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.622098 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0b9868b-191a-4a71-b538-e67f0dbaa54f-config\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.622163 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b693f37f-a924-4f57-a4a5-c9ba03815229-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.622353 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-auth-proxy-config\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.622794 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d8ce3e58-f4e8-444f-ba23-172619e34134-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.623167 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e1149c46-0323-453e-a042-4a6e8155364d-apiservice-cert\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.623339 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.624008 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01c58bf7-d620-4770-bfc1-b63c29cf32f9-trusted-ca\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.625622 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12ffe862-7453-46fe-8aae-95065549f876-metrics-tls\") pod \"dns-operator-744455d44c-lpf4d\" (UID: \"12ffe862-7453-46fe-8aae-95065549f876\") " pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.626603 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0b9868b-191a-4a71-b538-e67f0dbaa54f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.626708 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9a04339-599f-4f20-a2bf-9d12addcd11f-metrics-tls\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.627238 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8231653d-6197-4f1b-85bf-8f5c260d47fe-default-certificate\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.628463 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.622943 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a265d87b-7caf-4b02-8afe-388cef7069b2-trusted-ca\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.633442 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c58bf7-d620-4770-bfc1-b63c29cf32f9-serving-cert\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.633549 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/735074c0-8634-44aa-9cfe-cdd1a07633d4-srv-cert\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.636514 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e893d3c-7f15-44a6-b881-f658c0e06171-secret-volume\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.638073 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-92v7h\" (UID: \"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.638176 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0b9868b-191a-4a71-b538-e67f0dbaa54f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qzrws\" (UID: \"c0b9868b-191a-4a71-b538-e67f0dbaa54f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.651164 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c5c6f36-9c21-487e-85f3-633d81390939-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bhtkj\" (UID: \"2c5c6f36-9c21-487e-85f3-633d81390939\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.651844 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/735074c0-8634-44aa-9cfe-cdd1a07633d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.652346 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8ce3e58-f4e8-444f-ba23-172619e34134-serving-cert\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.656166 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pstf8\" (UniqueName: \"kubernetes.io/projected/2214172a-784b-4626-9e7d-e0a7a0bbfeb7-kube-api-access-pstf8\") pod \"package-server-manager-789f6589d5-gbcsj\" (UID: \"2214172a-784b-4626-9e7d-e0a7a0bbfeb7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.679604 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwn2k\" (UniqueName: \"kubernetes.io/projected/da2727c5-131f-460a-82a9-60e440d73aef-kube-api-access-bwn2k\") pod \"service-ca-operator-777779d784-74nhw\" (UID: \"da2727c5-131f-460a-82a9-60e440d73aef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.681570 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.682034 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.182019737 +0000 UTC m=+148.928352182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.694875 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8fr5\" (UniqueName: \"kubernetes.io/projected/674b6f22-7e92-44fa-9e8d-6e2085de4b3c-kube-api-access-d8fr5\") pod \"openshift-apiserver-operator-796bbdcf4f-rrhwk\" (UID: \"674b6f22-7e92-44fa-9e8d-6e2085de4b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.698779 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.707190 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl"] Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.716787 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.746768 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd64d\" (UniqueName: \"kubernetes.io/projected/5e893d3c-7f15-44a6-b881-f658c0e06171-kube-api-access-xd64d\") pod \"collect-profiles-29412825-drx52\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:17 crc kubenswrapper[4677]: W1203 13:49:17.753073 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80a1b9b2_cf93_4da9_8edb_e3b3197b5c5a.slice/crio-784596eac7cdc2e2d6c99676acf0ac5dd7d0828122053654378e909088c9523f WatchSource:0}: Error finding container 784596eac7cdc2e2d6c99676acf0ac5dd7d0828122053654378e909088c9523f: Status 404 returned error can't find the container with id 784596eac7cdc2e2d6c99676acf0ac5dd7d0828122053654378e909088c9523f Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.763834 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjn59\" (UniqueName: \"kubernetes.io/projected/20bba33b-274c-4099-ae84-4fc906ffc07b-kube-api-access-rjn59\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phhd\" (UID: \"20bba33b-274c-4099-ae84-4fc906ffc07b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.764697 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.777539 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z47xp\" (UniqueName: \"kubernetes.io/projected/12ffe862-7453-46fe-8aae-95065549f876-kube-api-access-z47xp\") pod \"dns-operator-744455d44c-lpf4d\" (UID: \"12ffe862-7453-46fe-8aae-95065549f876\") " pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.787539 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.787824 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.287812617 +0000 UTC m=+149.034145072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.805765 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.807727 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wmgk\" (UniqueName: \"kubernetes.io/projected/4f4d19d6-6b48-4c81-85d9-36a3c076d498-kube-api-access-2wmgk\") pod \"machine-config-server-r8sn4\" (UID: \"4f4d19d6-6b48-4c81-85d9-36a3c076d498\") " pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.813280 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhpl2\" (UniqueName: \"kubernetes.io/projected/894546c9-1858-4460-8a89-02cc798d7c4c-kube-api-access-mhpl2\") pod \"migrator-59844c95c7-l5kb9\" (UID: \"894546c9-1858-4460-8a89-02cc798d7c4c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.844319 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lztcv\" (UniqueName: \"kubernetes.io/projected/84259910-9b22-4957-ad88-7d00052ee57c-kube-api-access-lztcv\") pod \"route-controller-manager-6576b87f9c-vwsjj\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.848565 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.852716 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nbvm\" (UniqueName: \"kubernetes.io/projected/93bc7735-ccca-4199-b3b2-9bb1c2eeef7c-kube-api-access-8nbvm\") pod \"machine-approver-56656f9798-6m6rm\" (UID: \"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.859827 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.861112 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp5c6\" (UniqueName: \"kubernetes.io/projected/f9a04339-599f-4f20-a2bf-9d12addcd11f-kube-api-access-kp5c6\") pod \"dns-default-bq6q6\" (UID: \"f9a04339-599f-4f20-a2bf-9d12addcd11f\") " pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.870287 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" event={"ID":"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885","Type":"ContainerStarted","Data":"277aea29a37b5a377f8407b7b0b47c80a1ea3eabb6495ed534a4f25cfb3cd10b"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.872165 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"76d77c6b05d91816ea37604737deb1c2dc3c9086d3c068f0905873188066b61e"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.872187 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"11f107e3a42d6306e8f905b92494d3c7b51be8d48ba52b01d6c8d11af101545a"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.877719 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" event={"ID":"93a0fb86-4f19-4023-b777-c58381af5047","Type":"ContainerStarted","Data":"106a6e381abc84ebce497e1498501dae2fb8a10c69249d89750d643f4e1388cd"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.877984 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" event={"ID":"93a0fb86-4f19-4023-b777-c58381af5047","Type":"ContainerStarted","Data":"706b5c308cc1b15080a2aa9d18cfbc6ff85e1e1cdff084f6b8563061b4ce5a6e"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.877998 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" event={"ID":"93a0fb86-4f19-4023-b777-c58381af5047","Type":"ContainerStarted","Data":"9890f3f73eaafd68af6bd5254980900c666c7a16b7e04351e5047ff25f6a1832"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.878475 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqw5k\" (UniqueName: \"kubernetes.io/projected/58b0e34a-a6be-4a51-9ef4-4dd2234ec15f-kube-api-access-bqw5k\") pod \"multus-admission-controller-857f4d67dd-p4cw4\" (UID: \"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.883109 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.891569 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" event={"ID":"b22d27e8-6223-4370-9c02-ef73ec251631","Type":"ContainerStarted","Data":"3fd18eb1a7faa05f3873f341a4ddd55df8fa892b2255f0516d12a3b24b1bfdce"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.891618 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" event={"ID":"b22d27e8-6223-4370-9c02-ef73ec251631","Type":"ContainerStarted","Data":"efad9c818ccff235f6b00795f3679034b3617c6741a5b5b6be28a8c69d2f2445"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.892223 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.892669 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.392651388 +0000 UTC m=+149.138983843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.894856 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"77f4fdb210882382ce93a96f4c439232760b9c86141460caaa4be741856e0b51"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.894893 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7349669b074f39c98e17a437bf9b3034d16d14f140edba041c242f8b8670b878"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.898843 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.902939 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r47j4\" (UniqueName: \"kubernetes.io/projected/735074c0-8634-44aa-9cfe-cdd1a07633d4-kube-api-access-r47j4\") pod \"olm-operator-6b444d44fb-fq4ql\" (UID: \"735074c0-8634-44aa-9cfe-cdd1a07633d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.906732 4677 generic.go:334] "Generic (PLEG): container finished" podID="f7e33975-80e2-410a-8819-3d66e41ccd21" containerID="aa7388aa7ae69d70ab96362304841899a5614cb73c30b8a929bb899fdd0a83a9" exitCode=0 Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.906789 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" event={"ID":"f7e33975-80e2-410a-8819-3d66e41ccd21","Type":"ContainerDied","Data":"aa7388aa7ae69d70ab96362304841899a5614cb73c30b8a929bb899fdd0a83a9"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.906814 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" event={"ID":"f7e33975-80e2-410a-8819-3d66e41ccd21","Type":"ContainerStarted","Data":"39a9e2a47722a5e5547f8d89a42c1d2dcbc7af442ef3994c321c6760e750a5d4"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.909633 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88"] Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.911348 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" event={"ID":"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a","Type":"ContainerStarted","Data":"784596eac7cdc2e2d6c99676acf0ac5dd7d0828122053654378e909088c9523f"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.912792 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8d2ead330605b3340220513a1cd8f7d680a2699bd14fb454d4a66382919177af"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.912820 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2d0c63fcfcafd1ab5de369eaaf648a5973f665737aeffe6555c7891ef5d8bec5"} Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.913207 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.913473 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.923282 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f5t2\" (UniqueName: \"kubernetes.io/projected/01c58bf7-d620-4770-bfc1-b63c29cf32f9-kube-api-access-8f5t2\") pod \"console-operator-58897d9998-66lfj\" (UID: \"01c58bf7-d620-4770-bfc1-b63c29cf32f9\") " pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.936519 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.955499 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqzt7\" (UniqueName: \"kubernetes.io/projected/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-kube-api-access-gqzt7\") pod \"console-f9d7485db-s84pq\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.955726 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.986345 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.991975 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.994205 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:17 crc kubenswrapper[4677]: E1203 13:49:17.994628 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.494608426 +0000 UTC m=+149.240940881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:17 crc kubenswrapper[4677]: I1203 13:49:17.995029 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.001059 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.007725 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.031925 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-r8sn4" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.071154 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.071571 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.095181 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.095433 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.595403396 +0000 UTC m=+149.341735851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.095515 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.095823 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.595810449 +0000 UTC m=+149.342142904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.182565 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8jrz\" (UniqueName: \"kubernetes.io/projected/d8ce3e58-f4e8-444f-ba23-172619e34134-kube-api-access-m8jrz\") pod \"openshift-config-operator-7777fb866f-xxds8\" (UID: \"d8ce3e58-f4e8-444f-ba23-172619e34134\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.194180 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgbhr\" (UniqueName: \"kubernetes.io/projected/9c0a9578-40a5-4345-8a4b-3e809f978d48-kube-api-access-pgbhr\") pod \"downloads-7954f5f757-t9z7p\" (UID: \"9c0a9578-40a5-4345-8a4b-3e809f978d48\") " pod="openshift-console/downloads-7954f5f757-t9z7p" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.196363 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.196859 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.696839447 +0000 UTC m=+149.443171922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.201904 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmvks\" (UniqueName: \"kubernetes.io/projected/93eeaace-9fca-4e02-909a-3db0b7ab56e4-kube-api-access-fmvks\") pod \"machine-config-controller-84d6567774-2psbm\" (UID: \"93eeaace-9fca-4e02-909a-3db0b7ab56e4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.202986 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c84b5\" (UniqueName: \"kubernetes.io/projected/a265d87b-7caf-4b02-8afe-388cef7069b2-kube-api-access-c84b5\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.203285 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97wxc\" (UniqueName: \"kubernetes.io/projected/8231653d-6197-4f1b-85bf-8f5c260d47fe-kube-api-access-97wxc\") pod \"router-default-5444994796-hmrqp\" (UID: \"8231653d-6197-4f1b-85bf-8f5c260d47fe\") " pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.205732 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.211908 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qgn7\" (UniqueName: \"kubernetes.io/projected/b693f37f-a924-4f57-a4a5-c9ba03815229-kube-api-access-4qgn7\") pod \"authentication-operator-69f744f599-8qnws\" (UID: \"b693f37f-a924-4f57-a4a5-c9ba03815229\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.224634 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcqpg\" (UniqueName: \"kubernetes.io/projected/2c5c6f36-9c21-487e-85f3-633d81390939-kube-api-access-hcqpg\") pod \"cluster-samples-operator-665b6dd947-bhtkj\" (UID: \"2c5c6f36-9c21-487e-85f3-633d81390939\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.225352 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntqz8"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.226659 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4sgb\" (UniqueName: \"kubernetes.io/projected/9dc01e2f-388f-41a2-a139-5d251ad6cda8-kube-api-access-z4sgb\") pod \"marketplace-operator-79b997595-lqjhf\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.227084 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqnb4\" (UniqueName: \"kubernetes.io/projected/515d4b6a-61f4-4929-b3be-d152d8ced855-kube-api-access-vqnb4\") pod \"service-ca-9c57cc56f-xmbtm\" (UID: \"515d4b6a-61f4-4929-b3be-d152d8ced855\") " pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.228566 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2djz\" (UniqueName: \"kubernetes.io/projected/4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9-kube-api-access-r2djz\") pod \"ingress-canary-ltrs7\" (UID: \"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9\") " pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.228960 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.229262 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngfpz\" (UniqueName: \"kubernetes.io/projected/a40585ec-d5f1-4c47-99c0-9fcfeebf2812-kube-api-access-ngfpz\") pod \"kube-storage-version-migrator-operator-b67b599dd-mz6hw\" (UID: \"a40585ec-d5f1-4c47-99c0-9fcfeebf2812\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.229867 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk5pv\" (UniqueName: \"kubernetes.io/projected/e1149c46-0323-453e-a042-4a6e8155364d-kube-api-access-nk5pv\") pod \"packageserver-d55dfcdfc-bhbs4\" (UID: \"e1149c46-0323-453e-a042-4a6e8155364d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.232838 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbm7c\" (UniqueName: \"kubernetes.io/projected/8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41-kube-api-access-pbm7c\") pod \"catalog-operator-68c6474976-sxt9n\" (UID: \"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.239600 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a265d87b-7caf-4b02-8afe-388cef7069b2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-d8hkz\" (UID: \"a265d87b-7caf-4b02-8afe-388cef7069b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.256859 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ff07b87-8826-40b5-a881-9cdabc473090-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8nwnf\" (UID: \"8ff07b87-8826-40b5-a881-9cdabc473090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.267464 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.273532 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.276930 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t79qg\" (UniqueName: \"kubernetes.io/projected/dfeedbc7-41ac-4853-9254-7b8876109abf-kube-api-access-t79qg\") pod \"csi-hostpathplugin-ftj4m\" (UID: \"dfeedbc7-41ac-4853-9254-7b8876109abf\") " pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.281771 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.299418 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.299732 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.799719755 +0000 UTC m=+149.546052200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.308620 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.324338 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.331605 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.356967 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.377290 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ltrs7" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.382128 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.397854 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.400634 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.400808 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.900779514 +0000 UTC m=+149.647111969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.401020 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.401471 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:18.901450895 +0000 UTC m=+149.647783350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.463102 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9glj\" (UniqueName: \"kubernetes.io/projected/da040c14-6dff-4426-8258-6bd0e3e14cb6-kube-api-access-f9glj\") pod \"control-plane-machine-set-operator-78cbb6b69f-2c9gq\" (UID: \"da040c14-6dff-4426-8258-6bd0e3e14cb6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.468469 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-t9z7p" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.475749 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.507143 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.508757 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.008733265 +0000 UTC m=+149.755065720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.520044 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.546824 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.570871 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.578424 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.580425 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.587486 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-55xx9"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.609405 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.609663 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.109649159 +0000 UTC m=+149.855981614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.656234 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.660023 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.661710 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.673028 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9"] Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.712547 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.713005 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.212981191 +0000 UTC m=+149.959313646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.720978 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lpf4d"] Dec 03 13:49:18 crc kubenswrapper[4677]: W1203 13:49:18.756647 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod674b6f22_7e92_44fa_9e8d_6e2085de4b3c.slice/crio-6943f2a4f245136b8d875cae4f7af50f304613929c519a46b907f91ad3334099 WatchSource:0}: Error finding container 6943f2a4f245136b8d875cae4f7af50f304613929c519a46b907f91ad3334099: Status 404 returned error can't find the container with id 6943f2a4f245136b8d875cae4f7af50f304613929c519a46b907f91ad3334099 Dec 03 13:49:18 crc kubenswrapper[4677]: W1203 13:49:18.763216 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20bba33b_274c_4099_ae84_4fc906ffc07b.slice/crio-46503bb3187e1e81cb0ccc87724a2aeff96f523a5f556470a285383b6e837a24 WatchSource:0}: Error finding container 46503bb3187e1e81cb0ccc87724a2aeff96f523a5f556470a285383b6e837a24: Status 404 returned error can't find the container with id 46503bb3187e1e81cb0ccc87724a2aeff96f523a5f556470a285383b6e837a24 Dec 03 13:49:18 crc kubenswrapper[4677]: W1203 13:49:18.770779 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60a15680_2160_4cea_a7e0_0a3a22685b3f.slice/crio-a4f90c374bd85d97ba2865ccbed74b3285b51a152c389ffe1280abb45f968e8a WatchSource:0}: Error finding container a4f90c374bd85d97ba2865ccbed74b3285b51a152c389ffe1280abb45f968e8a: Status 404 returned error can't find the container with id a4f90c374bd85d97ba2865ccbed74b3285b51a152c389ffe1280abb45f968e8a Dec 03 13:49:18 crc kubenswrapper[4677]: W1203 13:49:18.773380 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b3f1a3e_c3aa_4799_8906_07a4bb9eef2c.slice/crio-8ae3e52b41cc31a74a88a4bd6be5aa89ed1ad11f29a7c3b71c61b31d03ab6495 WatchSource:0}: Error finding container 8ae3e52b41cc31a74a88a4bd6be5aa89ed1ad11f29a7c3b71c61b31d03ab6495: Status 404 returned error can't find the container with id 8ae3e52b41cc31a74a88a4bd6be5aa89ed1ad11f29a7c3b71c61b31d03ab6495 Dec 03 13:49:18 crc kubenswrapper[4677]: W1203 13:49:18.775641 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0b9868b_191a_4a71_b538_e67f0dbaa54f.slice/crio-6bfe58ff3c1c1bb4f0ef801cffe08097b15287d8a1e51a7b7a684a1e365f1ba5 WatchSource:0}: Error finding container 6bfe58ff3c1c1bb4f0ef801cffe08097b15287d8a1e51a7b7a684a1e365f1ba5: Status 404 returned error can't find the container with id 6bfe58ff3c1c1bb4f0ef801cffe08097b15287d8a1e51a7b7a684a1e365f1ba5 Dec 03 13:49:18 crc kubenswrapper[4677]: W1203 13:49:18.778080 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84259910_9b22_4957_ad88_7d00052ee57c.slice/crio-0aa3ba1806b4b598a7d990bf09893d1183aec1454270660f3d39698065315925 WatchSource:0}: Error finding container 0aa3ba1806b4b598a7d990bf09893d1183aec1454270660f3d39698065315925: Status 404 returned error can't find the container with id 0aa3ba1806b4b598a7d990bf09893d1183aec1454270660f3d39698065315925 Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.815689 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.815994 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.315980932 +0000 UTC m=+150.062313387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:18 crc kubenswrapper[4677]: I1203 13:49:18.916426 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:18 crc kubenswrapper[4677]: E1203 13:49:18.917122 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.417102264 +0000 UTC m=+150.163434729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.095977 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.096319 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.596307117 +0000 UTC m=+150.342639562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.181416 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-74nhw"] Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.198489 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52"] Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.198534 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-66lfj"] Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.198545 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm"] Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.199023 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.199503 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.699485094 +0000 UTC m=+150.445817549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.208881 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" event={"ID":"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c","Type":"ContainerStarted","Data":"8cd3e24eacfd10104e1dd36c08713f223e266013b0d37c4ea187eae5f2eeda44"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.208967 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql"] Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.219786 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" event={"ID":"674b6f22-7e92-44fa-9e8d-6e2085de4b3c","Type":"ContainerStarted","Data":"6943f2a4f245136b8d875cae4f7af50f304613929c519a46b907f91ad3334099"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.229929 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" event={"ID":"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c","Type":"ContainerStarted","Data":"8ae3e52b41cc31a74a88a4bd6be5aa89ed1ad11f29a7c3b71c61b31d03ab6495"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.239695 4677 generic.go:334] "Generic (PLEG): container finished" podID="3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885" containerID="7f98367e6ec3b4b8a40682adffbc06388b7d039ccf370fd79ff83d74e1e4baa1" exitCode=0 Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.239764 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" event={"ID":"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885","Type":"ContainerDied","Data":"7f98367e6ec3b4b8a40682adffbc06388b7d039ccf370fd79ff83d74e1e4baa1"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.243453 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" event={"ID":"60a15680-2160-4cea-a7e0-0a3a22685b3f","Type":"ContainerStarted","Data":"a4f90c374bd85d97ba2865ccbed74b3285b51a152c389ffe1280abb45f968e8a"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.244665 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" event={"ID":"d91c7be8-82e6-4da1-b483-67f45cac28ef","Type":"ContainerStarted","Data":"889f711cd76985593629d80e7d85c4b9d3e80c1e77035290fb4bfc0fd3caa686"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.245884 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" event={"ID":"12ffe862-7453-46fe-8aae-95065549f876","Type":"ContainerStarted","Data":"b393ac14fea63d56b99939a18f1dd3de6f030e7e31027af75e0f39acf815660f"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.247051 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hmrqp" event={"ID":"8231653d-6197-4f1b-85bf-8f5c260d47fe","Type":"ContainerStarted","Data":"fef6f579ba75a3bdc7c05589c33bf64e83fae7b28d896fc8c585671da2d5a2c8"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.248391 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" event={"ID":"20bba33b-274c-4099-ae84-4fc906ffc07b","Type":"ContainerStarted","Data":"46503bb3187e1e81cb0ccc87724a2aeff96f523a5f556470a285383b6e837a24"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.249509 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-r8sn4" event={"ID":"4f4d19d6-6b48-4c81-85d9-36a3c076d498","Type":"ContainerStarted","Data":"ae144bfa4461028e5ea77d6d48f62cb097e5d195ecf4539340fcaa08f3ff97e7"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.250401 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" event={"ID":"c0b9868b-191a-4a71-b538-e67f0dbaa54f","Type":"ContainerStarted","Data":"6bfe58ff3c1c1bb4f0ef801cffe08097b15287d8a1e51a7b7a684a1e365f1ba5"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.251330 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" event={"ID":"894546c9-1858-4460-8a89-02cc798d7c4c","Type":"ContainerStarted","Data":"5c8f8c76dffa8a5a1cb45d80a72d0adbcca9542e4999b4bf98fd67308b9652fb"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.251983 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" event={"ID":"c2c20cb8-d990-4756-ad31-56245b583279","Type":"ContainerStarted","Data":"ea200524ede1c3f37a843e83822a3c91bb4c3d8818526f955b2347845365d270"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.259545 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" event={"ID":"80a1b9b2-cf93-4da9-8edb-e3b3197b5c5a","Type":"ContainerStarted","Data":"8db29fb2c3a61498479c8ead3ec8725120759b6e3c76f5fabdf23be88f212875"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.271732 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" event={"ID":"84259910-9b22-4957-ad88-7d00052ee57c","Type":"ContainerStarted","Data":"0aa3ba1806b4b598a7d990bf09893d1183aec1454270660f3d39698065315925"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.274722 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" event={"ID":"2214172a-784b-4626-9e7d-e0a7a0bbfeb7","Type":"ContainerStarted","Data":"01b5ff1652fa4d87cb9e028c39a6b26d01c24f11398e48b2376db09cd0281fc2"} Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.275515 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.278649 4677 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-ppzxd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.278697 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" podUID="b22d27e8-6223-4370-9c02-ef73ec251631" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.300442 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.300812 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.800793421 +0000 UTC m=+150.547125896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.406641 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.407998 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:19.907912116 +0000 UTC m=+150.654244621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.429769 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ftj4m"] Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.466436 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s84pq"] Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.541938 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.545571 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.045548767 +0000 UTC m=+150.791881212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.649386 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.649583 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.149555831 +0000 UTC m=+150.895888286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.649859 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.650224 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.150212041 +0000 UTC m=+150.896544496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.750632 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.750813 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.250783515 +0000 UTC m=+150.997115980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.750882 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.751297 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.251286501 +0000 UTC m=+150.997619006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.853051 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.853478 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.353459646 +0000 UTC m=+151.099792101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.956339 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:19 crc kubenswrapper[4677]: E1203 13:49:19.956701 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.456689325 +0000 UTC m=+151.203021780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:19 crc kubenswrapper[4677]: I1203 13:49:19.993434 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" podStartSLOduration=129.993412642 podStartE2EDuration="2m9.993412642s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:19.961415786 +0000 UTC m=+150.707748261" watchObservedRunningTime="2025-12-03 13:49:19.993412642 +0000 UTC m=+150.739745107" Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.004807 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-d4p7j" podStartSLOduration=130.004784667 podStartE2EDuration="2m10.004784667s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:20.000661104 +0000 UTC m=+150.746993559" watchObservedRunningTime="2025-12-03 13:49:20.004784667 +0000 UTC m=+150.751117122" Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.065530 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqdkl" podStartSLOduration=130.062944931 podStartE2EDuration="2m10.062944931s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:20.062454084 +0000 UTC m=+150.808786549" watchObservedRunningTime="2025-12-03 13:49:20.062944931 +0000 UTC m=+150.809277386" Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.077210 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.078597 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.578581962 +0000 UTC m=+151.324914417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.181709 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.182052 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.682040258 +0000 UTC m=+151.428372703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.189547 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.194512 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-p4cw4"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.202096 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.282769 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.282927 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.78289459 +0000 UTC m=+151.529227055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.282997 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.283535 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.78352743 +0000 UTC m=+151.529859885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.314335 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" event={"ID":"c2c20cb8-d990-4756-ad31-56245b583279","Type":"ContainerStarted","Data":"e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.315146 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" event={"ID":"dfeedbc7-41ac-4853-9254-7b8876109abf","Type":"ContainerStarted","Data":"30f192fa161ae7812d4ab0d7273a8f882324ea56b8d554d08944f044525ed12e"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.315765 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" event={"ID":"5e893d3c-7f15-44a6-b881-f658c0e06171","Type":"ContainerStarted","Data":"26222ee820ec026ffe79033b4c75fc6b82ebdb009070d014648f5ec98b6bb3e8"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.316646 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-r8sn4" event={"ID":"4f4d19d6-6b48-4c81-85d9-36a3c076d498","Type":"ContainerStarted","Data":"fe893fb594a03b570d34b04ebc72b86cfb5a1336824928441b3fd0616edb06ef"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.317286 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s84pq" event={"ID":"58f5bde4-0a29-4a20-a332-e945cc8bc3a0","Type":"ContainerStarted","Data":"15b99d25c79a99f2acbd9ffd50d151e9d57bda6f611ca50c834f854ab0c8750a"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.318006 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" event={"ID":"93eeaace-9fca-4e02-909a-3db0b7ab56e4","Type":"ContainerStarted","Data":"0ab5398b90369f5b417308e8cbfae35700f0ae37abcbd68ee9fd08d4413e88df"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.318929 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" event={"ID":"da2727c5-131f-460a-82a9-60e440d73aef","Type":"ContainerStarted","Data":"0ded442896209ed296b9e2e1815e565a63f3a61d64680b98758b6782925b3e57"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.319969 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" event={"ID":"674b6f22-7e92-44fa-9e8d-6e2085de4b3c","Type":"ContainerStarted","Data":"4d77dcd5714e1bbc26595d31aa1a6e446bb189d20a97c6fb946522f53a72a157"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.321464 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" event={"ID":"735074c0-8634-44aa-9cfe-cdd1a07633d4","Type":"ContainerStarted","Data":"cff81794f0d2a4308656e434e70fc493a851be06d0cb1d056c08632ba21a75ce"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.322145 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-66lfj" event={"ID":"01c58bf7-d620-4770-bfc1-b63c29cf32f9","Type":"ContainerStarted","Data":"fb7996ef658ec09bfa18a8eeb6b4e3f796b6664bfd72991fe4ea1aeb4f90b350"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.323258 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" event={"ID":"f7e33975-80e2-410a-8819-3d66e41ccd21","Type":"ContainerStarted","Data":"a82fdb07829dfae5457744ba09061005ace100278ddfe7ab107baefda9fc19b2"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.339288 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" event={"ID":"2214172a-784b-4626-9e7d-e0a7a0bbfeb7","Type":"ContainerStarted","Data":"c426ca3ab5853d9bc5866bf64098bc120c6879564af37c8561d3d01a94828ee2"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.350504 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" event={"ID":"d91c7be8-82e6-4da1-b483-67f45cac28ef","Type":"ContainerStarted","Data":"c4a0d3068541a96dc943c443a5ba65eb3fe55237b2473a58bc7d32cbe7b2b5c3"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.363005 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bq6q6"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.364386 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" event={"ID":"84259910-9b22-4957-ad88-7d00052ee57c","Type":"ContainerStarted","Data":"95544b6169ec2ca29b515fe97accf1326d91c64bc3fef16856e54a394199b0d7"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.366112 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" event={"ID":"20bba33b-274c-4099-ae84-4fc906ffc07b","Type":"ContainerStarted","Data":"b9ef6e8b94f8fcb50aebec323c7358785d44840726e70f4cf106ff0bb7eeaa1e"} Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.367127 4677 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-ppzxd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.367173 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" podUID="b22d27e8-6223-4370-9c02-ef73ec251631" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.380400 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ltrs7"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.384092 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.384171 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.884151246 +0000 UTC m=+151.630483701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.384207 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.384617 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.88460986 +0000 UTC m=+151.630942315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.387144 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.402761 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8qnws"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.403744 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-xmbtm"] Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.469998 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rrhwk" podStartSLOduration=130.469977827 podStartE2EDuration="2m10.469977827s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:20.469587175 +0000 UTC m=+151.215919670" watchObservedRunningTime="2025-12-03 13:49:20.469977827 +0000 UTC m=+151.216310292" Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.485632 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.486257 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:20.986220088 +0000 UTC m=+151.732552573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.612007 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.612752 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.112721762 +0000 UTC m=+151.859054257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.714913 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.715219 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.215203097 +0000 UTC m=+151.961535552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.816252 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.816701 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.31667891 +0000 UTC m=+152.063011405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.917499 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.917745 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.417706468 +0000 UTC m=+152.164038963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:20 crc kubenswrapper[4677]: I1203 13:49:20.918064 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:20 crc kubenswrapper[4677]: E1203 13:49:20.918575 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.418558815 +0000 UTC m=+152.164891310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.019648 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.019916 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.519889433 +0000 UTC m=+152.266221888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.020034 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.020436 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.520418329 +0000 UTC m=+152.266750774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.075604 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq"] Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.087909 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4"] Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.090169 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-t9z7p"] Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.095645 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lqjhf"] Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.097081 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xxds8"] Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.099669 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj"] Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.106640 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf"] Dec 03 13:49:21 crc kubenswrapper[4677]: W1203 13:49:21.107211 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb693f37f_a924_4f57_a4a5_c9ba03815229.slice/crio-c279cc9496aa2d6f2df6670cb873dc473ccae8c88163861e662c4e2841a9e92d WatchSource:0}: Error finding container c279cc9496aa2d6f2df6670cb873dc473ccae8c88163861e662c4e2841a9e92d: Status 404 returned error can't find the container with id c279cc9496aa2d6f2df6670cb873dc473ccae8c88163861e662c4e2841a9e92d Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.121036 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.121505 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.621459658 +0000 UTC m=+152.367792113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: W1203 13:49:21.127080 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dc01e2f_388f_41a2_a139_5d251ad6cda8.slice/crio-bdab2684993e917569777b4e9fe89e67087b0f8c42b9660818543cd9631c95e0 WatchSource:0}: Error finding container bdab2684993e917569777b4e9fe89e67087b0f8c42b9660818543cd9631c95e0: Status 404 returned error can't find the container with id bdab2684993e917569777b4e9fe89e67087b0f8c42b9660818543cd9631c95e0 Dec 03 13:49:21 crc kubenswrapper[4677]: W1203 13:49:21.133722 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1149c46_0323_453e_a042_4a6e8155364d.slice/crio-af6a55fa41cb57674f2967689b74ed70ad8231b3d0e9c2db8c53bd75247e327a WatchSource:0}: Error finding container af6a55fa41cb57674f2967689b74ed70ad8231b3d0e9c2db8c53bd75247e327a: Status 404 returned error can't find the container with id af6a55fa41cb57674f2967689b74ed70ad8231b3d0e9c2db8c53bd75247e327a Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.229394 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.229804 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.72978562 +0000 UTC m=+152.476118075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.330448 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.331659 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.831626794 +0000 UTC m=+152.577959259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.418182 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" event={"ID":"93eeaace-9fca-4e02-909a-3db0b7ab56e4","Type":"ContainerStarted","Data":"768d5e9cb57aad2e8512d316f98ef9f01c7ed4b753bdf79b7a59ffef7ecab1bc"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.420246 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" event={"ID":"2c5c6f36-9c21-487e-85f3-633d81390939","Type":"ContainerStarted","Data":"1597a6c9c76b8c1391f03644ecbab78595a55fa231d03c5490674eae59d1b72f"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.433856 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.436425 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:21.936408503 +0000 UTC m=+152.682740958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.464290 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" event={"ID":"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c","Type":"ContainerStarted","Data":"a3d827e7cdf567a9abb1d7f7aa9f7fe1faff659dfbc227b008c01aeef834231c"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.467100 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s84pq" event={"ID":"58f5bde4-0a29-4a20-a332-e945cc8bc3a0","Type":"ContainerStarted","Data":"dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.538568 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.539198 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.039172587 +0000 UTC m=+152.785505032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.599813 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" event={"ID":"3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885","Type":"ContainerStarted","Data":"468ea7ccfd3a58ad7e6614baafc17abe2886887af5f6d6d4c9a1960bb3264a56"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.646940 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.652413 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.152395596 +0000 UTC m=+152.898728051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.714302 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" event={"ID":"b693f37f-a924-4f57-a4a5-c9ba03815229","Type":"ContainerStarted","Data":"c279cc9496aa2d6f2df6670cb873dc473ccae8c88163861e662c4e2841a9e92d"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.734955 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hmrqp" event={"ID":"8231653d-6197-4f1b-85bf-8f5c260d47fe","Type":"ContainerStarted","Data":"b6eefd50c870cf98fc201fcb2402f52e970e7a38788077e1060384d017f97d67"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.742077 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bq6q6" event={"ID":"f9a04339-599f-4f20-a2bf-9d12addcd11f","Type":"ContainerStarted","Data":"e076f35f690e1c16a92069b50c74d7a8e19794d2bafee844d5947173fba354b8"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.749815 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" event={"ID":"735074c0-8634-44aa-9cfe-cdd1a07633d4","Type":"ContainerStarted","Data":"7b4c7b499264c6947adad9a64f81e8722451f2c911dee244c04b872796868ffd"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.751544 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.765284 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.766593 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.266566916 +0000 UTC m=+153.012899371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.774993 4677 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fq4ql container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.775050 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" podUID="735074c0-8634-44aa-9cfe-cdd1a07633d4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.785118 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-s84pq" podStartSLOduration=131.785104859 podStartE2EDuration="2m11.785104859s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:21.537513404 +0000 UTC m=+152.283845889" watchObservedRunningTime="2025-12-03 13:49:21.785104859 +0000 UTC m=+152.531437314" Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.807862 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" event={"ID":"d91c7be8-82e6-4da1-b483-67f45cac28ef","Type":"ContainerStarted","Data":"ac7e86d58ef0ed5cb41f09487148fea878f1c72e789bfdee820d874ae25f0d37"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.866735 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.868541 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.368528364 +0000 UTC m=+153.114860819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.903783 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-t9z7p" event={"ID":"9c0a9578-40a5-4345-8a4b-3e809f978d48","Type":"ContainerStarted","Data":"4d950c97ad39e228a83e9042433676c604e2d9c4fbf1f77806a46ac37a4aa6d9"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.917818 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" event={"ID":"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41","Type":"ContainerStarted","Data":"acb118b575ec3393ab413552baf065a5b37f2587dbca032a57c971cbc6be3e76"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.922993 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" event={"ID":"1b3f1a3e-c3aa-4799-8906-07a4bb9eef2c","Type":"ContainerStarted","Data":"ce350a9af63ca81485b189134a81737417a5c608bb4ac48b2293fed5c0a2e689"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.927064 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-66lfj" event={"ID":"01c58bf7-d620-4770-bfc1-b63c29cf32f9","Type":"ContainerStarted","Data":"f65b0a6c841ff737aad9c00efbc41f513b64029a99fcb55456137e6c96aa0712"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.927819 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.949820 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ltrs7" event={"ID":"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9","Type":"ContainerStarted","Data":"d605f3fe2fd731497a76982039ca655c58516f3de491ea6f5bbb2edb23a613f5"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.950503 4677 patch_prober.go:28] interesting pod/console-operator-58897d9998-66lfj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.950560 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-66lfj" podUID="01c58bf7-d620-4770-bfc1-b63c29cf32f9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.952190 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" event={"ID":"c0b9868b-191a-4a71-b538-e67f0dbaa54f","Type":"ContainerStarted","Data":"9efc41520617d36938b037b2a8a787d74440ee084492f533dd0e753716404cf0"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.954221 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" event={"ID":"894546c9-1858-4460-8a89-02cc798d7c4c","Type":"ContainerStarted","Data":"99473a0a1cf6c4ae04fba7fdf533feda685ee7863a6a0d4873d2a50aa491a505"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.954978 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" event={"ID":"9dc01e2f-388f-41a2-a139-5d251ad6cda8","Type":"ContainerStarted","Data":"bdab2684993e917569777b4e9fe89e67087b0f8c42b9660818543cd9631c95e0"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.955780 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" event={"ID":"da040c14-6dff-4426-8258-6bd0e3e14cb6","Type":"ContainerStarted","Data":"364b42884c697b7b733599454e85067576ebe87d8555f310af4082a24b0dda8e"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.957222 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" event={"ID":"5e893d3c-7f15-44a6-b881-f658c0e06171","Type":"ContainerStarted","Data":"f7bdd33c43e30e88ef4fbe6b4ad11ae175d53bdb573475dbf86d0ae528cdd267"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.968571 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:21 crc kubenswrapper[4677]: E1203 13:49:21.970231 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.470179511 +0000 UTC m=+153.216511966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.997167 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" event={"ID":"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f","Type":"ContainerStarted","Data":"f8652c3e2d1f091af38343230dd852a971cc35753d1807fcaeb0e257de3d166e"} Dec 03 13:49:21 crc kubenswrapper[4677]: I1203 13:49:21.997204 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" event={"ID":"12ffe862-7453-46fe-8aae-95065549f876","Type":"ContainerStarted","Data":"9964a38adb57c50ffcd7d398bad4072f2718b9f7ea66b65d21d8a4e99d2c35f8"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.005113 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" podStartSLOduration=132.005093311 podStartE2EDuration="2m12.005093311s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:21.785728679 +0000 UTC m=+152.532061134" watchObservedRunningTime="2025-12-03 13:49:22.005093311 +0000 UTC m=+152.751425766" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.006939 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-hmrqp" podStartSLOduration=132.006931819 podStartE2EDuration="2m12.006931819s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.00289195 +0000 UTC m=+152.749224415" watchObservedRunningTime="2025-12-03 13:49:22.006931819 +0000 UTC m=+152.753264264" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.010497 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" event={"ID":"e1149c46-0323-453e-a042-4a6e8155364d","Type":"ContainerStarted","Data":"af6a55fa41cb57674f2967689b74ed70ad8231b3d0e9c2db8c53bd75247e327a"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.040486 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" event={"ID":"60a15680-2160-4cea-a7e0-0a3a22685b3f","Type":"ContainerStarted","Data":"38b95277037e453f344c86e0b5bcfb26c2454966095fd7968b56076b36a037c3"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.043690 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" event={"ID":"d8ce3e58-f4e8-444f-ba23-172619e34134","Type":"ContainerStarted","Data":"81d0e66f1c366502e4ebdf5ce858c56c5140c63e67d261c055ac09b91ddf5abd"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.045189 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" event={"ID":"a40585ec-d5f1-4c47-99c0-9fcfeebf2812","Type":"ContainerStarted","Data":"c54c6cdba51a92a167cb65b8ef66f7b863cf75b5b6aa94e66723bd2563c2e3f4"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.046672 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" event={"ID":"515d4b6a-61f4-4929-b3be-d152d8ced855","Type":"ContainerStarted","Data":"0f4334b6d147549b3b9bcfc30720c8930ceb33a6ff4bc71f91b1f88c5545a95a"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.047642 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" event={"ID":"a265d87b-7caf-4b02-8afe-388cef7069b2","Type":"ContainerStarted","Data":"027465327c2e7194795bebe1148b4aa703075ba4383bf388bed4b230ac14211f"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.048614 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" event={"ID":"da2727c5-131f-460a-82a9-60e440d73aef","Type":"ContainerStarted","Data":"8bd6902d0c325d2f45f7a5b907c9271b3c45770146e6472b5cfff1214f07d131"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.058764 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" podStartSLOduration=132.05874426 podStartE2EDuration="2m12.05874426s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.055030142 +0000 UTC m=+152.801362617" watchObservedRunningTime="2025-12-03 13:49:22.05874426 +0000 UTC m=+152.805076715" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.065785 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" event={"ID":"f7e33975-80e2-410a-8819-3d66e41ccd21","Type":"ContainerStarted","Data":"af7ebfc0b75a3eb5b83d9bbd81f162553706c778b5a6672df7ab4079681f7b82"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.068595 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" event={"ID":"8ff07b87-8826-40b5-a881-9cdabc473090","Type":"ContainerStarted","Data":"8d6c585e758350ad6ecf2b90741b1d57405eea54e02271580768a7f205d930e5"} Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.070458 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.072424 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.572400768 +0000 UTC m=+153.318733413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.103139 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmq88" podStartSLOduration=132.103118713 podStartE2EDuration="2m12.103118713s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.100916682 +0000 UTC m=+152.847249147" watchObservedRunningTime="2025-12-03 13:49:22.103118713 +0000 UTC m=+152.849451168" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.127240 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" podStartSLOduration=132.127225325 podStartE2EDuration="2m12.127225325s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.124641472 +0000 UTC m=+152.870973927" watchObservedRunningTime="2025-12-03 13:49:22.127225325 +0000 UTC m=+152.873557780" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.171522 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.172385 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.672342992 +0000 UTC m=+153.418675447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.177022 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-92v7h" podStartSLOduration=132.176090632 podStartE2EDuration="2m12.176090632s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.172211407 +0000 UTC m=+152.918543882" watchObservedRunningTime="2025-12-03 13:49:22.176090632 +0000 UTC m=+152.922423097" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.177315 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-74nhw" podStartSLOduration=131.17730719 podStartE2EDuration="2m11.17730719s" podCreationTimestamp="2025-12-03 13:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.149198119 +0000 UTC m=+152.895530574" watchObservedRunningTime="2025-12-03 13:49:22.17730719 +0000 UTC m=+152.923639645" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.192388 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qzrws" podStartSLOduration=132.192368814 podStartE2EDuration="2m12.192368814s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.19037684 +0000 UTC m=+152.936709295" watchObservedRunningTime="2025-12-03 13:49:22.192368814 +0000 UTC m=+152.938701269" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.211706 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.212090 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.212329 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.270096 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" podStartSLOduration=132.270075024 podStartE2EDuration="2m12.270075024s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.237600873 +0000 UTC m=+152.983933338" watchObservedRunningTime="2025-12-03 13:49:22.270075024 +0000 UTC m=+153.016407489" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.272787 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-55xx9" podStartSLOduration=132.2727756 podStartE2EDuration="2m12.2727756s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.269731833 +0000 UTC m=+153.016064298" watchObservedRunningTime="2025-12-03 13:49:22.2727756 +0000 UTC m=+153.019108055" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.278781 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.279415 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.779396323 +0000 UTC m=+153.525728778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.316435 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-66lfj" podStartSLOduration=132.316416359 podStartE2EDuration="2m12.316416359s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.315166799 +0000 UTC m=+153.061499264" watchObservedRunningTime="2025-12-03 13:49:22.316416359 +0000 UTC m=+153.062748814" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.371930 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" podStartSLOduration=132.371909578 podStartE2EDuration="2m12.371909578s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.367993392 +0000 UTC m=+153.114325857" watchObservedRunningTime="2025-12-03 13:49:22.371909578 +0000 UTC m=+153.118242033" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.383134 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.383209 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.383892 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.384112 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.884098179 +0000 UTC m=+153.630430624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.384287 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.384628 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.884613905 +0000 UTC m=+153.630946360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.406402 4677 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-z92ld container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.406548 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" podUID="3e8fd1eb-ad7f-4230-8805-c3ac3b8f2885" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.407766 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-r8sn4" podStartSLOduration=7.407750047 podStartE2EDuration="7.407750047s" podCreationTimestamp="2025-12-03 13:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.407042094 +0000 UTC m=+153.153374549" watchObservedRunningTime="2025-12-03 13:49:22.407750047 +0000 UTC m=+153.154082502" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.454312 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" podStartSLOduration=132.454296709 podStartE2EDuration="2m12.454296709s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.452802261 +0000 UTC m=+153.199134736" watchObservedRunningTime="2025-12-03 13:49:22.454296709 +0000 UTC m=+153.200629164" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.486311 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.486561 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:22.986547053 +0000 UTC m=+153.732879508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.505748 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" podStartSLOduration=131.505733148 podStartE2EDuration="2m11.505733148s" podCreationTimestamp="2025-12-03 13:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.486571004 +0000 UTC m=+153.232903479" watchObservedRunningTime="2025-12-03 13:49:22.505733148 +0000 UTC m=+153.252065603" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.551185 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phhd" podStartSLOduration=132.551170384 podStartE2EDuration="2m12.551170384s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:22.548484448 +0000 UTC m=+153.294816923" watchObservedRunningTime="2025-12-03 13:49:22.551170384 +0000 UTC m=+153.297502839" Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.590276 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.590645 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.090633489 +0000 UTC m=+153.836965944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.691687 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.692111 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.192093711 +0000 UTC m=+153.938426176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.793517 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.793910 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.293892583 +0000 UTC m=+154.040225088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:22 crc kubenswrapper[4677]: I1203 13:49:22.895502 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:22 crc kubenswrapper[4677]: E1203 13:49:22.895861 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.395840581 +0000 UTC m=+154.142173036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.000353 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.001303 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.501288771 +0000 UTC m=+154.247621226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.106512 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.106883 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.606863245 +0000 UTC m=+154.353195700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.108756 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" event={"ID":"8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41","Type":"ContainerStarted","Data":"9c2a32e36ed9dd046e3a0027633de674e4929a346877e0720817832b01666bab"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.108984 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.110781 4677 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-sxt9n container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.110827 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" podUID="8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.114774 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" event={"ID":"dfeedbc7-41ac-4853-9254-7b8876109abf","Type":"ContainerStarted","Data":"26c95998f1089c8e51a0389bcfd143c2d4222af1f464237feb5dfbe916272a00"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.116723 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mz6hw" event={"ID":"a40585ec-d5f1-4c47-99c0-9fcfeebf2812","Type":"ContainerStarted","Data":"bd70b2dba1cd1e309803fbd6255c17ffc3a4c24f9fde89c453fc49c55a594444"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.140273 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" event={"ID":"515d4b6a-61f4-4929-b3be-d152d8ced855","Type":"ContainerStarted","Data":"f9be34518bbca8c1895a8326da4bf2d19478023177997225354f3cad8cd76a40"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.141388 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" podStartSLOduration=133.141375421 podStartE2EDuration="2m13.141375421s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.129864682 +0000 UTC m=+153.876197147" watchObservedRunningTime="2025-12-03 13:49:23.141375421 +0000 UTC m=+153.887707876" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.156171 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" event={"ID":"b693f37f-a924-4f57-a4a5-c9ba03815229","Type":"ContainerStarted","Data":"6ba3ca97afad822993961f68bf293f4e3b74f1352c1ecb1c791419860785a3a2"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.168629 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" event={"ID":"a265d87b-7caf-4b02-8afe-388cef7069b2","Type":"ContainerStarted","Data":"2762df3cdd713c2001a2eca66b252e30a9ff4ca8f6bc0600d7e84fe20136bfd7"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.168683 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" event={"ID":"a265d87b-7caf-4b02-8afe-388cef7069b2","Type":"ContainerStarted","Data":"4830cc1bfbb6d5ef034d03203b4a8780540167cbe6e411ab4a5eaab854968807"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.177991 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" event={"ID":"da040c14-6dff-4426-8258-6bd0e3e14cb6","Type":"ContainerStarted","Data":"f31aee82a23fa2a4696bc7ad13f62b9395bdc6aa53e6e76d81814bb853809132"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.182015 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" event={"ID":"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f","Type":"ContainerStarted","Data":"d4c9d61cc43e9b37e7697aabf2594bb94ab4ab9583e9f2e5da9587689e97394e"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.186929 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" event={"ID":"894546c9-1858-4460-8a89-02cc798d7c4c","Type":"ContainerStarted","Data":"c13c5a7cbf812b9027362869b7138cb47853c30b28b3dc16d4b63176516a3a70"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.199698 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" event={"ID":"93eeaace-9fca-4e02-909a-3db0b7ab56e4","Type":"ContainerStarted","Data":"3185b568c4ee14ab74be595722c6cbb91ed8beb81c023095cb4a2da45413ccc3"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.207704 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.209027 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.70901351 +0000 UTC m=+154.455345975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.212647 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:23 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:23 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:23 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.212690 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.213211 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" event={"ID":"8ff07b87-8826-40b5-a881-9cdabc473090","Type":"ContainerStarted","Data":"58f10556366ec18061dc724e2cd58d7238be03959cd5f9c4e3588bd9347fb7f6"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.216281 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bq6q6" event={"ID":"f9a04339-599f-4f20-a2bf-9d12addcd11f","Type":"ContainerStarted","Data":"0cac6916518c879a67dd42ca40c35aae33bd77182f6cae254beb379b869bc470"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.222565 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" event={"ID":"e1149c46-0323-453e-a042-4a6e8155364d","Type":"ContainerStarted","Data":"a7d8b3927a45888e7a5bcf99131882196b9f2914a3f11d553e67e6d7973f6816"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.223680 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.230225 4677 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bhbs4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.230283 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" podUID="e1149c46-0323-453e-a042-4a6e8155364d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.231669 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" podStartSLOduration=133.231654475 podStartE2EDuration="2m13.231654475s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.227712308 +0000 UTC m=+153.974044763" watchObservedRunningTime="2025-12-03 13:49:23.231654475 +0000 UTC m=+153.977986930" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.233461 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-xmbtm" podStartSLOduration=132.233448932 podStartE2EDuration="2m12.233448932s" podCreationTimestamp="2025-12-03 13:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.176548259 +0000 UTC m=+153.922880814" watchObservedRunningTime="2025-12-03 13:49:23.233448932 +0000 UTC m=+153.979781387" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.234348 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" event={"ID":"9dc01e2f-388f-41a2-a139-5d251ad6cda8","Type":"ContainerStarted","Data":"9370c5745888934d7eaa524156e7b23861ec199211dc3d49521f028841ea3f51"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.234844 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.244993 4677 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-lqjhf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.245052 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.246849 4677 generic.go:334] "Generic (PLEG): container finished" podID="d8ce3e58-f4e8-444f-ba23-172619e34134" containerID="38270101c5f0107212159ff3e0f938eb87c425b098461134af31ab609144efc9" exitCode=0 Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.246923 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" event={"ID":"d8ce3e58-f4e8-444f-ba23-172619e34134","Type":"ContainerDied","Data":"38270101c5f0107212159ff3e0f938eb87c425b098461134af31ab609144efc9"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.264781 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-t9z7p" event={"ID":"9c0a9578-40a5-4345-8a4b-3e809f978d48","Type":"ContainerStarted","Data":"b3e71052dd15affb94ffbd3e6f92e75ad6c64681443430679eafcf71e986f28d"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.265831 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-t9z7p" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.279369 4677 patch_prober.go:28] interesting pod/downloads-7954f5f757-t9z7p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.279431 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t9z7p" podUID="9c0a9578-40a5-4345-8a4b-3e809f978d48" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.287718 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" event={"ID":"2c5c6f36-9c21-487e-85f3-633d81390939","Type":"ContainerStarted","Data":"997363b21bd56001b0537c7a7bd1b795b0dc140421546249babdbcc75146b7a7"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.289701 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" event={"ID":"2214172a-784b-4626-9e7d-e0a7a0bbfeb7","Type":"ContainerStarted","Data":"05160a372e0a56d2ace8324cf59f9919575c7230ad120cacbe3e3974126aa0f8"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.290388 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.311584 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.313390 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.813369764 +0000 UTC m=+154.559702229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.327775 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ltrs7" event={"ID":"4ce5f7cb-0fb1-4a6a-aa57-9bcec01ee7e9","Type":"ContainerStarted","Data":"9a80586a0f92d14807f25357f7fc690e991c34d08c9fb467e98673ce264f4d87"} Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.328168 4677 patch_prober.go:28] interesting pod/console-operator-58897d9998-66lfj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.328212 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-66lfj" podUID="01c58bf7-d620-4770-bfc1-b63c29cf32f9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.328683 4677 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fq4ql container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.329431 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" podUID="735074c0-8634-44aa-9cfe-cdd1a07633d4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.438643 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.459320 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" podStartSLOduration=133.459301771 podStartE2EDuration="2m13.459301771s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.458113223 +0000 UTC m=+154.204445698" watchObservedRunningTime="2025-12-03 13:49:23.459301771 +0000 UTC m=+154.205634226" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.459696 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d8hkz" podStartSLOduration=133.459689844 podStartE2EDuration="2m13.459689844s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.266819342 +0000 UTC m=+154.013151807" watchObservedRunningTime="2025-12-03 13:49:23.459689844 +0000 UTC m=+154.206022299" Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.469891 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:23.96987469 +0000 UTC m=+154.716207135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.556199 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.635421 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.135399806 +0000 UTC m=+154.881732261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.695098 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2c9gq" podStartSLOduration=133.695077899 podStartE2EDuration="2m13.695077899s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.626934615 +0000 UTC m=+154.373267070" watchObservedRunningTime="2025-12-03 13:49:23.695077899 +0000 UTC m=+154.441410354" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.707481 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.707763 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.207751325 +0000 UTC m=+154.954083770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.756261 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" podStartSLOduration=133.756243719 podStartE2EDuration="2m13.756243719s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.753914114 +0000 UTC m=+154.500246569" watchObservedRunningTime="2025-12-03 13:49:23.756243719 +0000 UTC m=+154.502576194" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.808505 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.808694 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.308653609 +0000 UTC m=+155.054986064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.809284 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.809692 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.309680092 +0000 UTC m=+155.056012547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.895907 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-t9z7p" podStartSLOduration=133.895888595 podStartE2EDuration="2m13.895888595s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.838414783 +0000 UTC m=+154.584747248" watchObservedRunningTime="2025-12-03 13:49:23.895888595 +0000 UTC m=+154.642221050" Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.911471 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.911731 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.411707242 +0000 UTC m=+155.158039697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.912176 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:23 crc kubenswrapper[4677]: E1203 13:49:23.912496 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.412482997 +0000 UTC m=+155.158815452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:23 crc kubenswrapper[4677]: I1203 13:49:23.959997 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l5kb9" podStartSLOduration=133.95997738 podStartE2EDuration="2m13.95997738s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:23.959805054 +0000 UTC m=+154.706137509" watchObservedRunningTime="2025-12-03 13:49:23.95997738 +0000 UTC m=+154.706309855" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.013694 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.013891 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.513865877 +0000 UTC m=+155.260198332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.013998 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.014327 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.514319951 +0000 UTC m=+155.260652406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.014558 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-ltrs7" podStartSLOduration=9.014540208 podStartE2EDuration="9.014540208s" podCreationTimestamp="2025-12-03 13:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.009503626 +0000 UTC m=+154.755836071" watchObservedRunningTime="2025-12-03 13:49:24.014540208 +0000 UTC m=+154.760872663" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.115353 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.115536 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.615509714 +0000 UTC m=+155.361842159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.115738 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.116072 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.616065182 +0000 UTC m=+155.362397637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.210446 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:24 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:24 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:24 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.210511 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.217087 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.217251 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.717229524 +0000 UTC m=+155.463561979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.217337 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.217670 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.717655069 +0000 UTC m=+155.463987524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.318466 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.318677 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.818641356 +0000 UTC m=+155.564973801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.318748 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.319130 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.819114551 +0000 UTC m=+155.565447006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.333347 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" event={"ID":"2c5c6f36-9c21-487e-85f3-633d81390939","Type":"ContainerStarted","Data":"9b1eb83361a4fc5d1a7e44b65976956391d2250499e4c18d637b2b806e999f86"} Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.335212 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" event={"ID":"58b0e34a-a6be-4a51-9ef4-4dd2234ec15f","Type":"ContainerStarted","Data":"1ce55a9a041088203e6e14d5c673d7eb19767c932398c2f6eefe4eebdffa97da"} Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.336817 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bq6q6" event={"ID":"f9a04339-599f-4f20-a2bf-9d12addcd11f","Type":"ContainerStarted","Data":"592d8c58fca80bfabe11d3c602e91fa34375f82952a9b448f6dc46024b1e6319"} Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.336967 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.338458 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" event={"ID":"93bc7735-ccca-4199-b3b2-9bb1c2eeef7c","Type":"ContainerStarted","Data":"9a4cd2c430ac1edf2056dbe6a56476f5a53961887bcbf1816e12472a79c67145"} Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.340198 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" event={"ID":"12ffe862-7453-46fe-8aae-95065549f876","Type":"ContainerStarted","Data":"e04a820f11b230e324cd07bc0fa736f5d1f3e912a5094dce0fa34dac024f8da2"} Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.342528 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" event={"ID":"d8ce3e58-f4e8-444f-ba23-172619e34134","Type":"ContainerStarted","Data":"d760682220dfdf6d431d9611f3b831cf101fbdf10a6d9537ee0dea0660087e3e"} Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.343384 4677 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-sxt9n container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.343427 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" podUID="8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.343469 4677 patch_prober.go:28] interesting pod/console-operator-58897d9998-66lfj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.343514 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-66lfj" podUID="01c58bf7-d620-4770-bfc1-b63c29cf32f9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.343896 4677 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-lqjhf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.343923 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.345314 4677 patch_prober.go:28] interesting pod/downloads-7954f5f757-t9z7p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.345778 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t9z7p" podUID="9c0a9578-40a5-4345-8a4b-3e809f978d48" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.382409 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.384130 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" podStartSLOduration=134.384107454 podStartE2EDuration="2m14.384107454s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.194096873 +0000 UTC m=+154.940429348" watchObservedRunningTime="2025-12-03 13:49:24.384107454 +0000 UTC m=+155.130439909" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.384724 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" podStartSLOduration=134.384718813 podStartE2EDuration="2m14.384718813s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.382149841 +0000 UTC m=+155.128482296" watchObservedRunningTime="2025-12-03 13:49:24.384718813 +0000 UTC m=+155.131051268" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.402024 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.419680 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.419863 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.919827009 +0000 UTC m=+155.666159464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.420738 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.424647 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:24.924634332 +0000 UTC m=+155.670966787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.469216 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2psbm" podStartSLOduration=134.46918474 podStartE2EDuration="2m14.46918474s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.466550326 +0000 UTC m=+155.212882781" watchObservedRunningTime="2025-12-03 13:49:24.46918474 +0000 UTC m=+155.215517195" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.522144 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.522432 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.022386135 +0000 UTC m=+155.768718590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.522731 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.523269 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.023259684 +0000 UTC m=+155.769592139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.623786 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.624140 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.124124427 +0000 UTC m=+155.870456882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.629714 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bq6q6" podStartSLOduration=9.629697996 podStartE2EDuration="9.629697996s" podCreationTimestamp="2025-12-03 13:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.613420174 +0000 UTC m=+155.359752619" watchObservedRunningTime="2025-12-03 13:49:24.629697996 +0000 UTC m=+155.376030441" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.630821 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-p4cw4" podStartSLOduration=134.630816721 podStartE2EDuration="2m14.630816721s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.529324198 +0000 UTC m=+155.275656653" watchObservedRunningTime="2025-12-03 13:49:24.630816721 +0000 UTC m=+155.377149176" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.640683 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8nwnf" podStartSLOduration=134.640650897 podStartE2EDuration="2m14.640650897s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.64014622 +0000 UTC m=+155.386478675" watchObservedRunningTime="2025-12-03 13:49:24.640650897 +0000 UTC m=+155.386983352" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.697727 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-lpf4d" podStartSLOduration=134.697707225 podStartE2EDuration="2m14.697707225s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.697193419 +0000 UTC m=+155.443525894" watchObservedRunningTime="2025-12-03 13:49:24.697707225 +0000 UTC m=+155.444039680" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.724915 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.725309 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.22529773 +0000 UTC m=+155.971630185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.732289 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" podStartSLOduration=134.732275713 podStartE2EDuration="2m14.732275713s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.731501128 +0000 UTC m=+155.477833583" watchObservedRunningTime="2025-12-03 13:49:24.732275713 +0000 UTC m=+155.478608168" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.778033 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6m6rm" podStartSLOduration=134.778009359 podStartE2EDuration="2m14.778009359s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:24.770692404 +0000 UTC m=+155.517024849" watchObservedRunningTime="2025-12-03 13:49:24.778009359 +0000 UTC m=+155.524341814" Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.827577 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.827776 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.327759633 +0000 UTC m=+156.074092088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:24 crc kubenswrapper[4677]: I1203 13:49:24.827911 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:24 crc kubenswrapper[4677]: E1203 13:49:24.828242 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.328236009 +0000 UTC m=+156.074568464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:24.932510 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:24.932870 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.432855942 +0000 UTC m=+156.179188387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.035581 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.036168 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.536149543 +0000 UTC m=+156.282481998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.145986 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.146197 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.64618327 +0000 UTC m=+156.392515725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.237503 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:25 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:25 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:25 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.237574 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.247109 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.247432 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.747417184 +0000 UTC m=+156.493749639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.343151 4677 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bhbs4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.343224 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" podUID="e1149c46-0323-453e-a042-4a6e8155364d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.347901 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.348112 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.848076331 +0000 UTC m=+156.594408786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.348310 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.348631 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.848615978 +0000 UTC m=+156.594948433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.348771 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" event={"ID":"dfeedbc7-41ac-4853-9254-7b8876109abf","Type":"ContainerStarted","Data":"6555b3016ce66959f9a7fc7a87e7526fb2d3e8f39b2e655436813c12044255da"} Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.350322 4677 patch_prober.go:28] interesting pod/downloads-7954f5f757-t9z7p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.350377 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t9z7p" podUID="9c0a9578-40a5-4345-8a4b-3e809f978d48" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.449859 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.450061 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.950031099 +0000 UTC m=+156.696363554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.450433 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.450703 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:25.95069037 +0000 UTC m=+156.697022825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.551056 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.551259 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.051228033 +0000 UTC m=+156.797560498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.553208 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.553569 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.053553146 +0000 UTC m=+156.799885601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.655624 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.656284 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.156268699 +0000 UTC m=+156.902601144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.757930 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.758227 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.258215207 +0000 UTC m=+157.004547662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.859438 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.859629 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.359604827 +0000 UTC m=+157.105937282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.859691 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.860000 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.359985379 +0000 UTC m=+157.106317834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.960520 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.960698 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.460671156 +0000 UTC m=+157.207003611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:25 crc kubenswrapper[4677]: I1203 13:49:25.961186 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:25 crc kubenswrapper[4677]: E1203 13:49:25.961518 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.461507523 +0000 UTC m=+157.207840038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.061986 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.062347 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.562332324 +0000 UTC m=+157.308664779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.163429 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.163701 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.663690123 +0000 UTC m=+157.410022578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.213605 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:26 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:26 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:26 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.213658 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.264912 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.265311 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.76529187 +0000 UTC m=+157.511624325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.297076 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2x9kl"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.298234 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.300616 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.310489 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2x9kl"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.349453 4677 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bhbs4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.349544 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" podUID="e1149c46-0323-453e-a042-4a6e8155364d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.356529 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" event={"ID":"dfeedbc7-41ac-4853-9254-7b8876109abf","Type":"ContainerStarted","Data":"235fe1ef0f7620cec06729a67d170ab4813e5d103b56fa09d38652a8b6b33cce"} Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.366455 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-utilities\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.366509 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-catalog-content\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.366536 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.366559 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdmbn\" (UniqueName: \"kubernetes.io/projected/4deaa535-2caf-4758-8470-19e0d3d2d6dd-kube-api-access-cdmbn\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.367214 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.867198436 +0000 UTC m=+157.613530891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.468095 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.468319 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.968288126 +0000 UTC m=+157.714620581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.468728 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-utilities\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.468770 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-catalog-content\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.468798 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.468829 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdmbn\" (UniqueName: \"kubernetes.io/projected/4deaa535-2caf-4758-8470-19e0d3d2d6dd-kube-api-access-cdmbn\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.469362 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-utilities\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.469418 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:26.969400162 +0000 UTC m=+157.715732617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.469602 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-catalog-content\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.480702 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pvzwn"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.481616 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.495342 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.496210 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.497971 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.505370 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.505574 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.506560 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdmbn\" (UniqueName: \"kubernetes.io/projected/4deaa535-2caf-4758-8470-19e0d3d2d6dd-kube-api-access-cdmbn\") pod \"community-operators-2x9kl\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.508551 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvzwn"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.525977 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.568272 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.569330 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.569822 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwgww\" (UniqueName: \"kubernetes.io/projected/73e747c5-106b-479c-9264-558bf7783d87-kube-api-access-xwgww\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.569869 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fea112e3-dd87-434a-aa17-3a15ddc0de13-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.569927 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-catalog-content\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.569972 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fea112e3-dd87-434a-aa17-3a15ddc0de13-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.569991 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-utilities\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.570089 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:27.070072438 +0000 UTC m=+157.816404893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.611210 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.672133 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwgww\" (UniqueName: \"kubernetes.io/projected/73e747c5-106b-479c-9264-558bf7783d87-kube-api-access-xwgww\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.672206 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fea112e3-dd87-434a-aa17-3a15ddc0de13-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.672263 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.672310 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-catalog-content\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.672349 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fea112e3-dd87-434a-aa17-3a15ddc0de13-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.672380 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-utilities\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.673373 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-utilities\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.673410 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:27.17339195 +0000 UTC m=+157.919724405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.674904 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fea112e3-dd87-434a-aa17-3a15ddc0de13-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.675241 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-catalog-content\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.707376 4677 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.717209 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vhghx"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.718487 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.729161 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vhghx"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.760110 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwgww\" (UniqueName: \"kubernetes.io/projected/73e747c5-106b-479c-9264-558bf7783d87-kube-api-access-xwgww\") pod \"certified-operators-pvzwn\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.772572 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fea112e3-dd87-434a-aa17-3a15ddc0de13-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.774492 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.774842 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:27.274823821 +0000 UTC m=+158.021156276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.795277 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.841184 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.878390 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stfmk\" (UniqueName: \"kubernetes.io/projected/33e127ec-407a-4911-9478-b54509612df8-kube-api-access-stfmk\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.878476 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.878516 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-utilities\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.878585 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-catalog-content\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.878905 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:27.378892638 +0000 UTC m=+158.125225093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.888467 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dbnmw"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.889342 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.912716 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbnmw"] Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.979911 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.980144 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-catalog-content\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.980168 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stfmk\" (UniqueName: \"kubernetes.io/projected/33e127ec-407a-4911-9478-b54509612df8-kube-api-access-stfmk\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.980223 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-utilities\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.980677 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-utilities\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: E1203 13:49:26.980742 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 13:49:27.480727781 +0000 UTC m=+158.227060236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.980967 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-catalog-content\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.995100 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:26 crc kubenswrapper[4677]: I1203 13:49:26.995490 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.025798 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stfmk\" (UniqueName: \"kubernetes.io/projected/33e127ec-407a-4911-9478-b54509612df8-kube-api-access-stfmk\") pod \"community-operators-vhghx\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.047566 4677 patch_prober.go:28] interesting pod/apiserver-76f77b778f-gd9tn container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]log ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]etcd ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/max-in-flight-filter ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 13:49:27 crc kubenswrapper[4677]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 13:49:27 crc kubenswrapper[4677]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/openshift.io-startinformers ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 13:49:27 crc kubenswrapper[4677]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 13:49:27 crc kubenswrapper[4677]: livez check failed Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.047657 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" podUID="f7e33975-80e2-410a-8819-3d66e41ccd21" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.048483 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.082136 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.082619 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mcns\" (UniqueName: \"kubernetes.io/projected/96689c77-5d8c-46c9-a87b-fc98bf5797ed-kube-api-access-5mcns\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.082656 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-utilities\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.082724 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-catalog-content\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: E1203 13:49:27.083115 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 13:49:27.583096372 +0000 UTC m=+158.329428837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hct2b" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.165145 4677 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T13:49:26.707418441Z","Handler":null,"Name":""} Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.173164 4677 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.173206 4677 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.183653 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.183827 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mcns\" (UniqueName: \"kubernetes.io/projected/96689c77-5d8c-46c9-a87b-fc98bf5797ed-kube-api-access-5mcns\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.183864 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-utilities\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.183975 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-catalog-content\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.184642 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-utilities\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.184695 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-catalog-content\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.210281 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.217562 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mcns\" (UniqueName: \"kubernetes.io/projected/96689c77-5d8c-46c9-a87b-fc98bf5797ed-kube-api-access-5mcns\") pod \"certified-operators-dbnmw\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.220229 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.223357 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:27 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:27 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:27 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.223406 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.285544 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.327270 4677 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.327308 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.384900 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.390095 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" event={"ID":"dfeedbc7-41ac-4853-9254-7b8876109abf","Type":"ContainerStarted","Data":"6241a179f33b5315aa807595d654b7b3674f024c87d55420b7838babecf2b0cb"} Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.408989 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.429096 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z92ld" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.457607 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2x9kl"] Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.672714 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" podStartSLOduration=12.67268954 podStartE2EDuration="12.67268954s" podCreationTimestamp="2025-12-03 13:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:27.600663601 +0000 UTC m=+158.346996076" watchObservedRunningTime="2025-12-03 13:49:27.67268954 +0000 UTC m=+158.419022015" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.676836 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvzwn"] Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.685816 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hct2b\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.698957 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.713883 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.737689 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.849198 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.856051 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.898103 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vhghx"] Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.912591 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbnmw"] Dec 03 13:49:27 crc kubenswrapper[4677]: W1203 13:49:27.915811 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33e127ec_407a_4911_9478_b54509612df8.slice/crio-e1bd24574cd3ffba918b7b68a5c63e459d4bf8a16cf1ad0d4aa59b50e6fdf978 WatchSource:0}: Error finding container e1bd24574cd3ffba918b7b68a5c63e459d4bf8a16cf1ad0d4aa59b50e6fdf978: Status 404 returned error can't find the container with id e1bd24574cd3ffba918b7b68a5c63e459d4bf8a16cf1ad0d4aa59b50e6fdf978 Dec 03 13:49:27 crc kubenswrapper[4677]: I1203 13:49:27.986370 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.008254 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.008982 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.009369 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.015160 4677 patch_prober.go:28] interesting pod/console-f9d7485db-s84pq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.015207 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-s84pq" podUID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.081425 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-66lfj" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.190674 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hct2b"] Dec 03 13:49:28 crc kubenswrapper[4677]: W1203 13:49:28.197972 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0963f4ad_95a1_4602_ab12_3073f3db0581.slice/crio-370351aa33c17df7adefca93a32ec225c1aca4c6dc1c19b51637f000fb0b04d1 WatchSource:0}: Error finding container 370351aa33c17df7adefca93a32ec225c1aca4c6dc1c19b51637f000fb0b04d1: Status 404 returned error can't find the container with id 370351aa33c17df7adefca93a32ec225c1aca4c6dc1c19b51637f000fb0b04d1 Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.210383 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.210630 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:28 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:28 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:28 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.210679 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.277463 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.283463 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vbdns"] Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.284399 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.288101 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.295546 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbdns"] Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.319811 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.377442 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.395939 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" event={"ID":"0963f4ad-95a1-4602-ab12-3073f3db0581","Type":"ContainerStarted","Data":"370351aa33c17df7adefca93a32ec225c1aca4c6dc1c19b51637f000fb0b04d1"} Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.397593 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvzwn" event={"ID":"73e747c5-106b-479c-9264-558bf7783d87","Type":"ContainerStarted","Data":"3ec2abab01f797b0643275babc158ef428eabe3ca5a9e83fb7498954f3f7ee1c"} Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.398671 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2x9kl" event={"ID":"4deaa535-2caf-4758-8470-19e0d3d2d6dd","Type":"ContainerStarted","Data":"c05ed6eca5c209971450d097770c8d666e150f618c2b489702ca452acd99ad72"} Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.400232 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhghx" event={"ID":"33e127ec-407a-4911-9478-b54509612df8","Type":"ContainerStarted","Data":"e1bd24574cd3ffba918b7b68a5c63e459d4bf8a16cf1ad0d4aa59b50e6fdf978"} Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.401062 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbnmw" event={"ID":"96689c77-5d8c-46c9-a87b-fc98bf5797ed","Type":"ContainerStarted","Data":"ee3d4cd03206327d2374fb88f1c248be17229b8981bb7286094232d5c5d8e928"} Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.402038 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fea112e3-dd87-434a-aa17-3a15ddc0de13","Type":"ContainerStarted","Data":"a3fcb698fc09c915434b313c735cadaccf9fdd8a6501df8a3ae1fde69f2717f3"} Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.416861 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-catalog-content\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.417084 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m67kw\" (UniqueName: \"kubernetes.io/projected/dab888c4-d0b5-47b5-b0b9-176672bde399-kube-api-access-m67kw\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.417311 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-utilities\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.470077 4677 patch_prober.go:28] interesting pod/downloads-7954f5f757-t9z7p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.470176 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t9z7p" podUID="9c0a9578-40a5-4345-8a4b-3e809f978d48" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.470092 4677 patch_prober.go:28] interesting pod/downloads-7954f5f757-t9z7p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.470262 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-t9z7p" podUID="9c0a9578-40a5-4345-8a4b-3e809f978d48" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.518822 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-utilities\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.518898 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-catalog-content\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.518989 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m67kw\" (UniqueName: \"kubernetes.io/projected/dab888c4-d0b5-47b5-b0b9-176672bde399-kube-api-access-m67kw\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.520608 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-catalog-content\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.520725 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-utilities\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.541036 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m67kw\" (UniqueName: \"kubernetes.io/projected/dab888c4-d0b5-47b5-b0b9-176672bde399-kube-api-access-m67kw\") pod \"redhat-marketplace-vbdns\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.599485 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.683401 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w4sxv"] Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.685211 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.693866 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4sxv"] Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.824117 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wgt\" (UniqueName: \"kubernetes.io/projected/04d74a24-b7f5-4026-bea7-7eb2caef99ba-kube-api-access-p7wgt\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.824231 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-utilities\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.824418 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-catalog-content\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.925811 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wgt\" (UniqueName: \"kubernetes.io/projected/04d74a24-b7f5-4026-bea7-7eb2caef99ba-kube-api-access-p7wgt\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.925875 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-utilities\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.926437 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-utilities\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.926843 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-catalog-content\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.926539 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-catalog-content\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:28 crc kubenswrapper[4677]: I1203 13:49:28.946373 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wgt\" (UniqueName: \"kubernetes.io/projected/04d74a24-b7f5-4026-bea7-7eb2caef99ba-kube-api-access-p7wgt\") pod \"redhat-marketplace-w4sxv\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.063066 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.086250 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbdns"] Dec 03 13:49:29 crc kubenswrapper[4677]: W1203 13:49:29.098361 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddab888c4_d0b5_47b5_b0b9_176672bde399.slice/crio-707b724a9411b864293ca5acacf8cd87e7e12207ad9a2958ae8c249411319cdf WatchSource:0}: Error finding container 707b724a9411b864293ca5acacf8cd87e7e12207ad9a2958ae8c249411319cdf: Status 404 returned error can't find the container with id 707b724a9411b864293ca5acacf8cd87e7e12207ad9a2958ae8c249411319cdf Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.210228 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:29 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:29 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:29 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.210274 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.301904 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4sxv"] Dec 03 13:49:29 crc kubenswrapper[4677]: W1203 13:49:29.375368 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04d74a24_b7f5_4026_bea7_7eb2caef99ba.slice/crio-a2ead7f0896779c1fd73a8cabad6366b7326565034ecea1dfd9e397628fb351b WatchSource:0}: Error finding container a2ead7f0896779c1fd73a8cabad6366b7326565034ecea1dfd9e397628fb351b: Status 404 returned error can't find the container with id a2ead7f0896779c1fd73a8cabad6366b7326565034ecea1dfd9e397628fb351b Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.430097 4677 generic.go:334] "Generic (PLEG): container finished" podID="fea112e3-dd87-434a-aa17-3a15ddc0de13" containerID="4e53005f8c6736bc60e7179830ee4abd3eb6bff866919fb976df29a33009e648" exitCode=0 Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.430169 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fea112e3-dd87-434a-aa17-3a15ddc0de13","Type":"ContainerDied","Data":"4e53005f8c6736bc60e7179830ee4abd3eb6bff866919fb976df29a33009e648"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.432854 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" event={"ID":"0963f4ad-95a1-4602-ab12-3073f3db0581","Type":"ContainerStarted","Data":"e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.432910 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.433936 4677 generic.go:334] "Generic (PLEG): container finished" podID="73e747c5-106b-479c-9264-558bf7783d87" containerID="9262255b00bdb2b4906c3c557f398012d0b88a4f46059c5cc04cbed27c988b36" exitCode=0 Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.433995 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvzwn" event={"ID":"73e747c5-106b-479c-9264-558bf7783d87","Type":"ContainerDied","Data":"9262255b00bdb2b4906c3c557f398012d0b88a4f46059c5cc04cbed27c988b36"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.435244 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.436003 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbdns" event={"ID":"dab888c4-d0b5-47b5-b0b9-176672bde399","Type":"ContainerStarted","Data":"0d5e42547fe6a3d30f7695c616007c8c59b36c66444d011114f13c0058fd77d7"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.436037 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbdns" event={"ID":"dab888c4-d0b5-47b5-b0b9-176672bde399","Type":"ContainerStarted","Data":"707b724a9411b864293ca5acacf8cd87e7e12207ad9a2958ae8c249411319cdf"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.437808 4677 generic.go:334] "Generic (PLEG): container finished" podID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerID="ae21a2713ae1e81f38be1fee15332a7ba25790d9b474a670714307062d256486" exitCode=0 Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.437871 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2x9kl" event={"ID":"4deaa535-2caf-4758-8470-19e0d3d2d6dd","Type":"ContainerDied","Data":"ae21a2713ae1e81f38be1fee15332a7ba25790d9b474a670714307062d256486"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.443775 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4sxv" event={"ID":"04d74a24-b7f5-4026-bea7-7eb2caef99ba","Type":"ContainerStarted","Data":"a2ead7f0896779c1fd73a8cabad6366b7326565034ecea1dfd9e397628fb351b"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.446496 4677 generic.go:334] "Generic (PLEG): container finished" podID="33e127ec-407a-4911-9478-b54509612df8" containerID="57e4287dc21ec268ee9bd6a3782bdb166b8c9099b8bd585b23347da72a30e1d8" exitCode=0 Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.447190 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhghx" event={"ID":"33e127ec-407a-4911-9478-b54509612df8","Type":"ContainerDied","Data":"57e4287dc21ec268ee9bd6a3782bdb166b8c9099b8bd585b23347da72a30e1d8"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.450517 4677 generic.go:334] "Generic (PLEG): container finished" podID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerID="f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb" exitCode=0 Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.450555 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbnmw" event={"ID":"96689c77-5d8c-46c9-a87b-fc98bf5797ed","Type":"ContainerDied","Data":"f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb"} Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.485008 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-th96p"] Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.486089 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.490748 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.492605 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-th96p"] Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.618023 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" podStartSLOduration=139.618005522 podStartE2EDuration="2m19.618005522s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:29.61546846 +0000 UTC m=+160.361800935" watchObservedRunningTime="2025-12-03 13:49:29.618005522 +0000 UTC m=+160.364337987" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.638298 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-catalog-content\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.638360 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2nbl\" (UniqueName: \"kubernetes.io/projected/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-kube-api-access-c2nbl\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.638453 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-utilities\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.739362 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-utilities\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.739436 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-catalog-content\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.739464 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2nbl\" (UniqueName: \"kubernetes.io/projected/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-kube-api-access-c2nbl\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.739877 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-utilities\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.741379 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-catalog-content\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.759520 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2nbl\" (UniqueName: \"kubernetes.io/projected/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-kube-api-access-c2nbl\") pod \"redhat-operators-th96p\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.881919 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2qhnw"] Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.883216 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.891295 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2qhnw"] Dec 03 13:49:29 crc kubenswrapper[4677]: I1203 13:49:29.898048 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.043707 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-catalog-content\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.044252 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn4th\" (UniqueName: \"kubernetes.io/projected/d9a210ab-e032-4834-bb1c-b9835456d960-kube-api-access-wn4th\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.044303 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-utilities\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.099787 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-th96p"] Dec 03 13:49:30 crc kubenswrapper[4677]: W1203 13:49:30.104183 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfde6c082_8a6f_4e3b_bf24_68be4bfdcf87.slice/crio-a9da1fa851c20e7fb00885ac62be74ad2da3e290fd5dfec888449b453e963a62 WatchSource:0}: Error finding container a9da1fa851c20e7fb00885ac62be74ad2da3e290fd5dfec888449b453e963a62: Status 404 returned error can't find the container with id a9da1fa851c20e7fb00885ac62be74ad2da3e290fd5dfec888449b453e963a62 Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.145250 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn4th\" (UniqueName: \"kubernetes.io/projected/d9a210ab-e032-4834-bb1c-b9835456d960-kube-api-access-wn4th\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.145301 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-utilities\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.145907 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-catalog-content\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.147684 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-catalog-content\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.148141 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-utilities\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.165217 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn4th\" (UniqueName: \"kubernetes.io/projected/d9a210ab-e032-4834-bb1c-b9835456d960-kube-api-access-wn4th\") pod \"redhat-operators-2qhnw\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.198299 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.209990 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:30 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:30 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:30 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.210075 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.419259 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2qhnw"] Dec 03 13:49:30 crc kubenswrapper[4677]: W1203 13:49:30.429956 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9a210ab_e032_4834_bb1c_b9835456d960.slice/crio-6b8d92e5ed8e3e06524576d7098915ef9faf7d171530803d672edd3e79c43ab8 WatchSource:0}: Error finding container 6b8d92e5ed8e3e06524576d7098915ef9faf7d171530803d672edd3e79c43ab8: Status 404 returned error can't find the container with id 6b8d92e5ed8e3e06524576d7098915ef9faf7d171530803d672edd3e79c43ab8 Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.464640 4677 generic.go:334] "Generic (PLEG): container finished" podID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerID="0d5e42547fe6a3d30f7695c616007c8c59b36c66444d011114f13c0058fd77d7" exitCode=0 Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.464704 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbdns" event={"ID":"dab888c4-d0b5-47b5-b0b9-176672bde399","Type":"ContainerDied","Data":"0d5e42547fe6a3d30f7695c616007c8c59b36c66444d011114f13c0058fd77d7"} Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.473373 4677 generic.go:334] "Generic (PLEG): container finished" podID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerID="52fdfa64d3010fcae3d5a8ea0b4564c38dce6d3fd68fbc163b2fcccf59ee827a" exitCode=0 Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.473474 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4sxv" event={"ID":"04d74a24-b7f5-4026-bea7-7eb2caef99ba","Type":"ContainerDied","Data":"52fdfa64d3010fcae3d5a8ea0b4564c38dce6d3fd68fbc163b2fcccf59ee827a"} Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.476673 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhnw" event={"ID":"d9a210ab-e032-4834-bb1c-b9835456d960","Type":"ContainerStarted","Data":"6b8d92e5ed8e3e06524576d7098915ef9faf7d171530803d672edd3e79c43ab8"} Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.478650 4677 generic.go:334] "Generic (PLEG): container finished" podID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerID="ab96a988e526fbbcdc2854cdbee810777a6c8a992507911101a5f6ac0a1d034c" exitCode=0 Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.478690 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th96p" event={"ID":"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87","Type":"ContainerDied","Data":"ab96a988e526fbbcdc2854cdbee810777a6c8a992507911101a5f6ac0a1d034c"} Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.478712 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th96p" event={"ID":"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87","Type":"ContainerStarted","Data":"a9da1fa851c20e7fb00885ac62be74ad2da3e290fd5dfec888449b453e963a62"} Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.500112 4677 generic.go:334] "Generic (PLEG): container finished" podID="5e893d3c-7f15-44a6-b881-f658c0e06171" containerID="f7bdd33c43e30e88ef4fbe6b4ad11ae175d53bdb573475dbf86d0ae528cdd267" exitCode=0 Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.500693 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" event={"ID":"5e893d3c-7f15-44a6-b881-f658c0e06171","Type":"ContainerDied","Data":"f7bdd33c43e30e88ef4fbe6b4ad11ae175d53bdb573475dbf86d0ae528cdd267"} Dec 03 13:49:30 crc kubenswrapper[4677]: I1203 13:49:30.947434 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.058312 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fea112e3-dd87-434a-aa17-3a15ddc0de13-kubelet-dir\") pod \"fea112e3-dd87-434a-aa17-3a15ddc0de13\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.058411 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fea112e3-dd87-434a-aa17-3a15ddc0de13-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fea112e3-dd87-434a-aa17-3a15ddc0de13" (UID: "fea112e3-dd87-434a-aa17-3a15ddc0de13"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.058979 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fea112e3-dd87-434a-aa17-3a15ddc0de13-kube-api-access\") pod \"fea112e3-dd87-434a-aa17-3a15ddc0de13\" (UID: \"fea112e3-dd87-434a-aa17-3a15ddc0de13\") " Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.063249 4677 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fea112e3-dd87-434a-aa17-3a15ddc0de13-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.067211 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea112e3-dd87-434a-aa17-3a15ddc0de13-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fea112e3-dd87-434a-aa17-3a15ddc0de13" (UID: "fea112e3-dd87-434a-aa17-3a15ddc0de13"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.146749 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 13:49:31 crc kubenswrapper[4677]: E1203 13:49:31.193899 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea112e3-dd87-434a-aa17-3a15ddc0de13" containerName="pruner" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.193978 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea112e3-dd87-434a-aa17-3a15ddc0de13" containerName="pruner" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.194499 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fea112e3-dd87-434a-aa17-3a15ddc0de13-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.194766 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="fea112e3-dd87-434a-aa17-3a15ddc0de13" containerName="pruner" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.195170 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.195243 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.202804 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.203020 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.210125 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:31 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:31 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:31 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.210295 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.399250 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0914e450-cb9c-4510-90bb-c238939ee065-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.399803 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0914e450-cb9c-4510-90bb-c238939ee065-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.502001 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0914e450-cb9c-4510-90bb-c238939ee065-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.502092 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0914e450-cb9c-4510-90bb-c238939ee065-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.502799 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0914e450-cb9c-4510-90bb-c238939ee065-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.511692 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fea112e3-dd87-434a-aa17-3a15ddc0de13","Type":"ContainerDied","Data":"a3fcb698fc09c915434b313c735cadaccf9fdd8a6501df8a3ae1fde69f2717f3"} Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.511747 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3fcb698fc09c915434b313c735cadaccf9fdd8a6501df8a3ae1fde69f2717f3" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.511835 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.533124 4677 generic.go:334] "Generic (PLEG): container finished" podID="d9a210ab-e032-4834-bb1c-b9835456d960" containerID="e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb" exitCode=0 Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.534251 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhnw" event={"ID":"d9a210ab-e032-4834-bb1c-b9835456d960","Type":"ContainerDied","Data":"e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb"} Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.550743 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0914e450-cb9c-4510-90bb-c238939ee065-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.828216 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.982477 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:31 crc kubenswrapper[4677]: I1203 13:49:31.999457 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.014750 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-gd9tn" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.027671 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd64d\" (UniqueName: \"kubernetes.io/projected/5e893d3c-7f15-44a6-b881-f658c0e06171-kube-api-access-xd64d\") pod \"5e893d3c-7f15-44a6-b881-f658c0e06171\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.027750 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e893d3c-7f15-44a6-b881-f658c0e06171-secret-volume\") pod \"5e893d3c-7f15-44a6-b881-f658c0e06171\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.027824 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e893d3c-7f15-44a6-b881-f658c0e06171-config-volume\") pod \"5e893d3c-7f15-44a6-b881-f658c0e06171\" (UID: \"5e893d3c-7f15-44a6-b881-f658c0e06171\") " Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.030581 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e893d3c-7f15-44a6-b881-f658c0e06171-config-volume" (OuterVolumeSpecName: "config-volume") pod "5e893d3c-7f15-44a6-b881-f658c0e06171" (UID: "5e893d3c-7f15-44a6-b881-f658c0e06171"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.037295 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e893d3c-7f15-44a6-b881-f658c0e06171-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5e893d3c-7f15-44a6-b881-f658c0e06171" (UID: "5e893d3c-7f15-44a6-b881-f658c0e06171"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.041539 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e893d3c-7f15-44a6-b881-f658c0e06171-kube-api-access-xd64d" (OuterVolumeSpecName: "kube-api-access-xd64d") pod "5e893d3c-7f15-44a6-b881-f658c0e06171" (UID: "5e893d3c-7f15-44a6-b881-f658c0e06171"). InnerVolumeSpecName "kube-api-access-xd64d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.131598 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e893d3c-7f15-44a6-b881-f658c0e06171-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.131640 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd64d\" (UniqueName: \"kubernetes.io/projected/5e893d3c-7f15-44a6-b881-f658c0e06171-kube-api-access-xd64d\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.131654 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e893d3c-7f15-44a6-b881-f658c0e06171-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.211242 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:32 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:32 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:32 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.211330 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.496939 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.560976 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" event={"ID":"5e893d3c-7f15-44a6-b881-f658c0e06171","Type":"ContainerDied","Data":"26222ee820ec026ffe79033b4c75fc6b82ebdb009070d014648f5ec98b6bb3e8"} Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.561009 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26222ee820ec026ffe79033b4c75fc6b82ebdb009070d014648f5ec98b6bb3e8" Dec 03 13:49:32 crc kubenswrapper[4677]: I1203 13:49:32.561582 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52" Dec 03 13:49:33 crc kubenswrapper[4677]: I1203 13:49:33.041576 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:49:33 crc kubenswrapper[4677]: I1203 13:49:33.063622 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d26430e4-fe9f-4b2f-ae90-a91fd8fccf79-metrics-certs\") pod \"network-metrics-daemon-j7h62\" (UID: \"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79\") " pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:49:33 crc kubenswrapper[4677]: I1203 13:49:33.074439 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bq6q6" Dec 03 13:49:33 crc kubenswrapper[4677]: I1203 13:49:33.230983 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:33 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:33 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:33 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:33 crc kubenswrapper[4677]: I1203 13:49:33.231071 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:33 crc kubenswrapper[4677]: I1203 13:49:33.298417 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j7h62" Dec 03 13:49:33 crc kubenswrapper[4677]: I1203 13:49:33.566833 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0914e450-cb9c-4510-90bb-c238939ee065","Type":"ContainerStarted","Data":"d9a39e7a8e024d967eb0615c718d67d68a7c16111588f960f0617c7dcab4b970"} Dec 03 13:49:34 crc kubenswrapper[4677]: I1203 13:49:34.065796 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-j7h62"] Dec 03 13:49:34 crc kubenswrapper[4677]: I1203 13:49:34.209298 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:34 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:34 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:34 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:34 crc kubenswrapper[4677]: I1203 13:49:34.209605 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:34 crc kubenswrapper[4677]: I1203 13:49:34.673606 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-j7h62" event={"ID":"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79","Type":"ContainerStarted","Data":"5d513d0eb78ff19323a5f146f8c17ec7ac4b403f7c7bf6766ffc76ef4b048b08"} Dec 03 13:49:34 crc kubenswrapper[4677]: I1203 13:49:34.677341 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0914e450-cb9c-4510-90bb-c238939ee065","Type":"ContainerStarted","Data":"e5dfb6fa719362872e72a5cf14f33886796f54c3c6e680752c59efdd0582ed14"} Dec 03 13:49:35 crc kubenswrapper[4677]: I1203 13:49:35.241469 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:35 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:35 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:35 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:35 crc kubenswrapper[4677]: I1203 13:49:35.241518 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:35 crc kubenswrapper[4677]: I1203 13:49:35.708913 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.708895039 podStartE2EDuration="4.708895039s" podCreationTimestamp="2025-12-03 13:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:35.706714329 +0000 UTC m=+166.453046804" watchObservedRunningTime="2025-12-03 13:49:35.708895039 +0000 UTC m=+166.455227494" Dec 03 13:49:36 crc kubenswrapper[4677]: I1203 13:49:36.208904 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:36 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:36 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:36 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:36 crc kubenswrapper[4677]: I1203 13:49:36.208976 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:36 crc kubenswrapper[4677]: I1203 13:49:36.693353 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-j7h62" event={"ID":"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79","Type":"ContainerStarted","Data":"6151d1fa37b6839587b6bb68a044d8b99bd75d8518d9f270590acf2b20ecdede"} Dec 03 13:49:37 crc kubenswrapper[4677]: I1203 13:49:37.210536 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:37 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:37 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:37 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:37 crc kubenswrapper[4677]: I1203 13:49:37.210582 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:37 crc kubenswrapper[4677]: I1203 13:49:37.717096 4677 generic.go:334] "Generic (PLEG): container finished" podID="0914e450-cb9c-4510-90bb-c238939ee065" containerID="e5dfb6fa719362872e72a5cf14f33886796f54c3c6e680752c59efdd0582ed14" exitCode=0 Dec 03 13:49:37 crc kubenswrapper[4677]: I1203 13:49:37.717533 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0914e450-cb9c-4510-90bb-c238939ee065","Type":"ContainerDied","Data":"e5dfb6fa719362872e72a5cf14f33886796f54c3c6e680752c59efdd0582ed14"} Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.009477 4677 patch_prober.go:28] interesting pod/console-f9d7485db-s84pq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.009851 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-s84pq" podUID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.214627 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:38 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:38 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:38 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.214726 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.437807 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.437871 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.470702 4677 patch_prober.go:28] interesting pod/downloads-7954f5f757-t9z7p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.470939 4677 patch_prober.go:28] interesting pod/downloads-7954f5f757-t9z7p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.471001 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t9z7p" podUID="9c0a9578-40a5-4345-8a4b-3e809f978d48" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 03 13:49:38 crc kubenswrapper[4677]: I1203 13:49:38.471000 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-t9z7p" podUID="9c0a9578-40a5-4345-8a4b-3e809f978d48" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.35:8080/\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.017388 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.177800 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0914e450-cb9c-4510-90bb-c238939ee065-kube-api-access\") pod \"0914e450-cb9c-4510-90bb-c238939ee065\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.177852 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0914e450-cb9c-4510-90bb-c238939ee065-kubelet-dir\") pod \"0914e450-cb9c-4510-90bb-c238939ee065\" (UID: \"0914e450-cb9c-4510-90bb-c238939ee065\") " Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.178312 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0914e450-cb9c-4510-90bb-c238939ee065-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0914e450-cb9c-4510-90bb-c238939ee065" (UID: "0914e450-cb9c-4510-90bb-c238939ee065"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.191707 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0914e450-cb9c-4510-90bb-c238939ee065-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0914e450-cb9c-4510-90bb-c238939ee065" (UID: "0914e450-cb9c-4510-90bb-c238939ee065"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.219553 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:39 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:39 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:39 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.219611 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.278908 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0914e450-cb9c-4510-90bb-c238939ee065-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.278960 4677 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0914e450-cb9c-4510-90bb-c238939ee065-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.742051 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0914e450-cb9c-4510-90bb-c238939ee065","Type":"ContainerDied","Data":"d9a39e7a8e024d967eb0615c718d67d68a7c16111588f960f0617c7dcab4b970"} Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.742088 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9a39e7a8e024d967eb0615c718d67d68a7c16111588f960f0617c7dcab4b970" Dec 03 13:49:39 crc kubenswrapper[4677]: I1203 13:49:39.742120 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 13:49:40 crc kubenswrapper[4677]: I1203 13:49:40.209550 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:40 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:40 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:40 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:40 crc kubenswrapper[4677]: I1203 13:49:40.209615 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:40 crc kubenswrapper[4677]: I1203 13:49:40.750739 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-j7h62" event={"ID":"d26430e4-fe9f-4b2f-ae90-a91fd8fccf79","Type":"ContainerStarted","Data":"7ffab6c75c67cbe6b8fc5df94f037ac1ff3da650843e1ef40d946d701c842023"} Dec 03 13:49:41 crc kubenswrapper[4677]: I1203 13:49:41.208886 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:41 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:41 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:41 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:41 crc kubenswrapper[4677]: I1203 13:49:41.209201 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:41 crc kubenswrapper[4677]: I1203 13:49:41.771665 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-j7h62" podStartSLOduration=151.771630122 podStartE2EDuration="2m31.771630122s" podCreationTimestamp="2025-12-03 13:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:49:41.770287579 +0000 UTC m=+172.516620044" watchObservedRunningTime="2025-12-03 13:49:41.771630122 +0000 UTC m=+172.517962577" Dec 03 13:49:42 crc kubenswrapper[4677]: I1203 13:49:42.208768 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 13:49:42 crc kubenswrapper[4677]: [-]has-synced failed: reason withheld Dec 03 13:49:42 crc kubenswrapper[4677]: [+]process-running ok Dec 03 13:49:42 crc kubenswrapper[4677]: healthz check failed Dec 03 13:49:42 crc kubenswrapper[4677]: I1203 13:49:42.208863 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 13:49:43 crc kubenswrapper[4677]: I1203 13:49:43.210630 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:43 crc kubenswrapper[4677]: I1203 13:49:43.216782 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-hmrqp" Dec 03 13:49:47 crc kubenswrapper[4677]: I1203 13:49:47.216992 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 13:49:47 crc kubenswrapper[4677]: I1203 13:49:47.720287 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:49:48 crc kubenswrapper[4677]: I1203 13:49:48.016986 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:48 crc kubenswrapper[4677]: I1203 13:49:48.025757 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 13:49:48 crc kubenswrapper[4677]: I1203 13:49:48.476839 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-t9z7p" Dec 03 13:49:57 crc kubenswrapper[4677]: I1203 13:49:57.725730 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" Dec 03 13:49:59 crc kubenswrapper[4677]: I1203 13:49:59.407200 4677 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-8qnws container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 13:49:59 crc kubenswrapper[4677]: I1203 13:49:59.407699 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" podUID="b693f37f-a924-4f57-a4a5-c9ba03815229" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.344419 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 13:50:08 crc kubenswrapper[4677]: E1203 13:50:08.345312 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0914e450-cb9c-4510-90bb-c238939ee065" containerName="pruner" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.345342 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0914e450-cb9c-4510-90bb-c238939ee065" containerName="pruner" Dec 03 13:50:08 crc kubenswrapper[4677]: E1203 13:50:08.345363 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e893d3c-7f15-44a6-b881-f658c0e06171" containerName="collect-profiles" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.345379 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e893d3c-7f15-44a6-b881-f658c0e06171" containerName="collect-profiles" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.345608 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0914e450-cb9c-4510-90bb-c238939ee065" containerName="pruner" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.345639 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e893d3c-7f15-44a6-b881-f658c0e06171" containerName="collect-profiles" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.346479 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.349920 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.351441 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.353720 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.399619 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f965bc9-867b-4abe-aca7-8a39f1863311-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.399881 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f965bc9-867b-4abe-aca7-8a39f1863311-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.437709 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.437795 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.500683 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f965bc9-867b-4abe-aca7-8a39f1863311-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.500779 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f965bc9-867b-4abe-aca7-8a39f1863311-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.500800 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f965bc9-867b-4abe-aca7-8a39f1863311-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.534097 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f965bc9-867b-4abe-aca7-8a39f1863311-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:08 crc kubenswrapper[4677]: I1203 13:50:08.679800 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.139452 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.140583 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.149560 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.317307 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.317485 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/536f457b-8005-409a-84e8-967f9d6f968d-kube-api-access\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.317531 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-var-lock\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.418474 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-var-lock\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.418591 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-var-lock\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.418635 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.418600 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.418710 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/536f457b-8005-409a-84e8-967f9d6f968d-kube-api-access\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.437473 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/536f457b-8005-409a-84e8-967f9d6f968d-kube-api-access\") pod \"installer-9-crc\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:13 crc kubenswrapper[4677]: I1203 13:50:13.498523 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:50:14 crc kubenswrapper[4677]: E1203 13:50:14.329910 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 13:50:14 crc kubenswrapper[4677]: E1203 13:50:14.330480 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cdmbn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-2x9kl_openshift-marketplace(4deaa535-2caf-4758-8470-19e0d3d2d6dd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:14 crc kubenswrapper[4677]: E1203 13:50:14.331740 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-2x9kl" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" Dec 03 13:50:15 crc kubenswrapper[4677]: E1203 13:50:15.801717 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 13:50:15 crc kubenswrapper[4677]: E1203 13:50:15.801874 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stfmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vhghx_openshift-marketplace(33e127ec-407a-4911-9478-b54509612df8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:15 crc kubenswrapper[4677]: E1203 13:50:15.803294 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vhghx" podUID="33e127ec-407a-4911-9478-b54509612df8" Dec 03 13:50:18 crc kubenswrapper[4677]: E1203 13:50:18.700739 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-2x9kl" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" Dec 03 13:50:18 crc kubenswrapper[4677]: E1203 13:50:18.701227 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vhghx" podUID="33e127ec-407a-4911-9478-b54509612df8" Dec 03 13:50:20 crc kubenswrapper[4677]: E1203 13:50:20.520112 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 13:50:20 crc kubenswrapper[4677]: E1203 13:50:20.520591 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m67kw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vbdns_openshift-marketplace(dab888c4-d0b5-47b5-b0b9-176672bde399): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:20 crc kubenswrapper[4677]: E1203 13:50:20.521770 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vbdns" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" Dec 03 13:50:27 crc kubenswrapper[4677]: E1203 13:50:27.918781 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vbdns" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.230887 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.231649 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p7wgt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-w4sxv_openshift-marketplace(04d74a24-b7f5-4026-bea7-7eb2caef99ba): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.233229 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-w4sxv" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.240972 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.241128 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wn4th,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2qhnw_openshift-marketplace(d9a210ab-e032-4834-bb1c-b9835456d960): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.242533 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2qhnw" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.359457 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.359893 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c2nbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-th96p_openshift-marketplace(fde6c082-8a6f-4e3b-bf24-68be4bfdcf87): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:31 crc kubenswrapper[4677]: E1203 13:50:31.361256 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-th96p" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.755606 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-th96p" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.755634 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-w4sxv" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.755651 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2qhnw" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.849985 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.850521 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5mcns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dbnmw_openshift-marketplace(96689c77-5d8c-46c9-a87b-fc98bf5797ed): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.852136 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dbnmw" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.860091 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.860262 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xwgww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-pvzwn_openshift-marketplace(73e747c5-106b-479c-9264-558bf7783d87): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 13:50:32 crc kubenswrapper[4677]: E1203 13:50:32.861783 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-pvzwn" podUID="73e747c5-106b-479c-9264-558bf7783d87" Dec 03 13:50:33 crc kubenswrapper[4677]: I1203 13:50:33.197919 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 13:50:33 crc kubenswrapper[4677]: I1203 13:50:33.202366 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 13:50:33 crc kubenswrapper[4677]: E1203 13:50:33.206181 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dbnmw" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" Dec 03 13:50:33 crc kubenswrapper[4677]: E1203 13:50:33.208337 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-pvzwn" podUID="73e747c5-106b-479c-9264-558bf7783d87" Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.211498 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5f965bc9-867b-4abe-aca7-8a39f1863311","Type":"ContainerStarted","Data":"e8e1cd57b93ac8c1289a16492e5f1a89e27ca936e05d403fc066e2aa0f1979b1"} Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.213102 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5f965bc9-867b-4abe-aca7-8a39f1863311","Type":"ContainerStarted","Data":"dad8973a1c9a482580c2e3cfe20d1470246e90a415e16be071fa5e8ce77cefb3"} Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.214655 4677 generic.go:334] "Generic (PLEG): container finished" podID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerID="15675fd847f4cead6f2819d806eb6587f4a4055bfaed31ada9f67f90279c1edb" exitCode=0 Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.214737 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2x9kl" event={"ID":"4deaa535-2caf-4758-8470-19e0d3d2d6dd","Type":"ContainerDied","Data":"15675fd847f4cead6f2819d806eb6587f4a4055bfaed31ada9f67f90279c1edb"} Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.216424 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"536f457b-8005-409a-84e8-967f9d6f968d","Type":"ContainerStarted","Data":"e29148d7b9ff52704c1a5ef775f2af31ab5c35747bea6896dcf1c8d075180ab1"} Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.216461 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"536f457b-8005-409a-84e8-967f9d6f968d","Type":"ContainerStarted","Data":"dbdf09959deaeaa84548ac87826dc8634bbcc397b450c1ae310025e9234ab1b0"} Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.232441 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=26.232422583 podStartE2EDuration="26.232422583s" podCreationTimestamp="2025-12-03 13:50:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:50:34.22954079 +0000 UTC m=+224.975873255" watchObservedRunningTime="2025-12-03 13:50:34.232422583 +0000 UTC m=+224.978755038" Dec 03 13:50:34 crc kubenswrapper[4677]: I1203 13:50:34.260793 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=21.26077608 podStartE2EDuration="21.26077608s" podCreationTimestamp="2025-12-03 13:50:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:50:34.257923898 +0000 UTC m=+225.004256363" watchObservedRunningTime="2025-12-03 13:50:34.26077608 +0000 UTC m=+225.007108535" Dec 03 13:50:35 crc kubenswrapper[4677]: I1203 13:50:35.223793 4677 generic.go:334] "Generic (PLEG): container finished" podID="5f965bc9-867b-4abe-aca7-8a39f1863311" containerID="e8e1cd57b93ac8c1289a16492e5f1a89e27ca936e05d403fc066e2aa0f1979b1" exitCode=0 Dec 03 13:50:35 crc kubenswrapper[4677]: I1203 13:50:35.224094 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5f965bc9-867b-4abe-aca7-8a39f1863311","Type":"ContainerDied","Data":"e8e1cd57b93ac8c1289a16492e5f1a89e27ca936e05d403fc066e2aa0f1979b1"} Dec 03 13:50:35 crc kubenswrapper[4677]: I1203 13:50:35.226729 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2x9kl" event={"ID":"4deaa535-2caf-4758-8470-19e0d3d2d6dd","Type":"ContainerStarted","Data":"208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a"} Dec 03 13:50:35 crc kubenswrapper[4677]: I1203 13:50:35.228874 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhghx" event={"ID":"33e127ec-407a-4911-9478-b54509612df8","Type":"ContainerStarted","Data":"07c2c6bbe4e24779b55d60ba31e5e01aa12fa36e6be1c2748ca1a73527284617"} Dec 03 13:50:35 crc kubenswrapper[4677]: I1203 13:50:35.282233 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2x9kl" podStartSLOduration=3.799739775 podStartE2EDuration="1m9.282204058s" podCreationTimestamp="2025-12-03 13:49:26 +0000 UTC" firstStartedPulling="2025-12-03 13:49:29.438899571 +0000 UTC m=+160.185232026" lastFinishedPulling="2025-12-03 13:50:34.921363854 +0000 UTC m=+225.667696309" observedRunningTime="2025-12-03 13:50:35.275640628 +0000 UTC m=+226.021973083" watchObservedRunningTime="2025-12-03 13:50:35.282204058 +0000 UTC m=+226.028536553" Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.235043 4677 generic.go:334] "Generic (PLEG): container finished" podID="33e127ec-407a-4911-9478-b54509612df8" containerID="07c2c6bbe4e24779b55d60ba31e5e01aa12fa36e6be1c2748ca1a73527284617" exitCode=0 Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.235128 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhghx" event={"ID":"33e127ec-407a-4911-9478-b54509612df8","Type":"ContainerDied","Data":"07c2c6bbe4e24779b55d60ba31e5e01aa12fa36e6be1c2748ca1a73527284617"} Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.495303 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.612049 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.612305 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f965bc9-867b-4abe-aca7-8a39f1863311-kubelet-dir\") pod \"5f965bc9-867b-4abe-aca7-8a39f1863311\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.612380 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f965bc9-867b-4abe-aca7-8a39f1863311-kube-api-access\") pod \"5f965bc9-867b-4abe-aca7-8a39f1863311\" (UID: \"5f965bc9-867b-4abe-aca7-8a39f1863311\") " Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.612389 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.612527 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f965bc9-867b-4abe-aca7-8a39f1863311-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5f965bc9-867b-4abe-aca7-8a39f1863311" (UID: "5f965bc9-867b-4abe-aca7-8a39f1863311"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.612698 4677 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f965bc9-867b-4abe-aca7-8a39f1863311-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.624063 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f965bc9-867b-4abe-aca7-8a39f1863311-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5f965bc9-867b-4abe-aca7-8a39f1863311" (UID: "5f965bc9-867b-4abe-aca7-8a39f1863311"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:50:36 crc kubenswrapper[4677]: I1203 13:50:36.714260 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f965bc9-867b-4abe-aca7-8a39f1863311-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:37 crc kubenswrapper[4677]: I1203 13:50:37.242379 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5f965bc9-867b-4abe-aca7-8a39f1863311","Type":"ContainerDied","Data":"dad8973a1c9a482580c2e3cfe20d1470246e90a415e16be071fa5e8ce77cefb3"} Dec 03 13:50:37 crc kubenswrapper[4677]: I1203 13:50:37.242423 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 13:50:37 crc kubenswrapper[4677]: I1203 13:50:37.242437 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dad8973a1c9a482580c2e3cfe20d1470246e90a415e16be071fa5e8ce77cefb3" Dec 03 13:50:38 crc kubenswrapper[4677]: I1203 13:50:38.140000 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-2x9kl" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="registry-server" probeResult="failure" output=< Dec 03 13:50:38 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 13:50:38 crc kubenswrapper[4677]: > Dec 03 13:50:38 crc kubenswrapper[4677]: I1203 13:50:38.436896 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:50:38 crc kubenswrapper[4677]: I1203 13:50:38.437023 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:50:38 crc kubenswrapper[4677]: I1203 13:50:38.437084 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:50:38 crc kubenswrapper[4677]: I1203 13:50:38.438253 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:50:38 crc kubenswrapper[4677]: I1203 13:50:38.438390 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4" gracePeriod=600 Dec 03 13:50:39 crc kubenswrapper[4677]: I1203 13:50:39.259142 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4" exitCode=0 Dec 03 13:50:39 crc kubenswrapper[4677]: I1203 13:50:39.259208 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4"} Dec 03 13:50:42 crc kubenswrapper[4677]: I1203 13:50:42.279042 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhghx" event={"ID":"33e127ec-407a-4911-9478-b54509612df8","Type":"ContainerStarted","Data":"11443ad15a7bd1a62b97fe0807f0d22985824db963597d82132adb8f0d375113"} Dec 03 13:50:42 crc kubenswrapper[4677]: I1203 13:50:42.281546 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"58bddf0714a1025ce6f4421667e6b66749f342b2705111136662a9104f1b7753"} Dec 03 13:50:42 crc kubenswrapper[4677]: I1203 13:50:42.299041 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vhghx" podStartSLOduration=4.187916615 podStartE2EDuration="1m16.299024105s" podCreationTimestamp="2025-12-03 13:49:26 +0000 UTC" firstStartedPulling="2025-12-03 13:49:29.448163058 +0000 UTC m=+160.194495513" lastFinishedPulling="2025-12-03 13:50:41.559270548 +0000 UTC m=+232.305603003" observedRunningTime="2025-12-03 13:50:42.296306518 +0000 UTC m=+233.042639003" watchObservedRunningTime="2025-12-03 13:50:42.299024105 +0000 UTC m=+233.045356560" Dec 03 13:50:43 crc kubenswrapper[4677]: I1203 13:50:43.288170 4677 generic.go:334] "Generic (PLEG): container finished" podID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerID="d39f678c854e6e145ee1f2ae7514cf96d300db7c527afa8ed27af30bb4a07c1b" exitCode=0 Dec 03 13:50:43 crc kubenswrapper[4677]: I1203 13:50:43.288248 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbdns" event={"ID":"dab888c4-d0b5-47b5-b0b9-176672bde399","Type":"ContainerDied","Data":"d39f678c854e6e145ee1f2ae7514cf96d300db7c527afa8ed27af30bb4a07c1b"} Dec 03 13:50:44 crc kubenswrapper[4677]: I1203 13:50:44.296274 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbdns" event={"ID":"dab888c4-d0b5-47b5-b0b9-176672bde399","Type":"ContainerStarted","Data":"3c205eea1e80fbb53e0d6a7cf27e6155be57a527c41b414b37c01dfe79a4e113"} Dec 03 13:50:44 crc kubenswrapper[4677]: I1203 13:50:44.301616 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th96p" event={"ID":"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87","Type":"ContainerStarted","Data":"1a9bca2aae97b40230bfa8ca39db60a2d916a15de70d02b0414411ce20def3b2"} Dec 03 13:50:44 crc kubenswrapper[4677]: I1203 13:50:44.333508 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vbdns" podStartSLOduration=2.778434848 podStartE2EDuration="1m16.333486372s" podCreationTimestamp="2025-12-03 13:49:28 +0000 UTC" firstStartedPulling="2025-12-03 13:49:30.473998718 +0000 UTC m=+161.220331173" lastFinishedPulling="2025-12-03 13:50:44.029050242 +0000 UTC m=+234.775382697" observedRunningTime="2025-12-03 13:50:44.315322701 +0000 UTC m=+235.061655166" watchObservedRunningTime="2025-12-03 13:50:44.333486372 +0000 UTC m=+235.079818847" Dec 03 13:50:45 crc kubenswrapper[4677]: I1203 13:50:45.309565 4677 generic.go:334] "Generic (PLEG): container finished" podID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerID="1a9bca2aae97b40230bfa8ca39db60a2d916a15de70d02b0414411ce20def3b2" exitCode=0 Dec 03 13:50:45 crc kubenswrapper[4677]: I1203 13:50:45.310150 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th96p" event={"ID":"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87","Type":"ContainerDied","Data":"1a9bca2aae97b40230bfa8ca39db60a2d916a15de70d02b0414411ce20def3b2"} Dec 03 13:50:46 crc kubenswrapper[4677]: I1203 13:50:46.655472 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:50:46 crc kubenswrapper[4677]: I1203 13:50:46.694286 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:50:47 crc kubenswrapper[4677]: I1203 13:50:47.050190 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:50:47 crc kubenswrapper[4677]: I1203 13:50:47.050252 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:50:47 crc kubenswrapper[4677]: I1203 13:50:47.097325 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:50:47 crc kubenswrapper[4677]: I1203 13:50:47.384494 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:50:48 crc kubenswrapper[4677]: I1203 13:50:48.600563 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:50:48 crc kubenswrapper[4677]: I1203 13:50:48.601703 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:50:48 crc kubenswrapper[4677]: I1203 13:50:48.650842 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:50:49 crc kubenswrapper[4677]: I1203 13:50:49.315818 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vhghx"] Dec 03 13:50:49 crc kubenswrapper[4677]: I1203 13:50:49.332104 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4sxv" event={"ID":"04d74a24-b7f5-4026-bea7-7eb2caef99ba","Type":"ContainerStarted","Data":"4e701f20f53836c75b78740cabc3e29706f3ad17bb3ad377069c9b305edd4113"} Dec 03 13:50:49 crc kubenswrapper[4677]: I1203 13:50:49.332514 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vhghx" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="registry-server" containerID="cri-o://11443ad15a7bd1a62b97fe0807f0d22985824db963597d82132adb8f0d375113" gracePeriod=2 Dec 03 13:50:49 crc kubenswrapper[4677]: I1203 13:50:49.380872 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:50:50 crc kubenswrapper[4677]: I1203 13:50:50.345164 4677 generic.go:334] "Generic (PLEG): container finished" podID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerID="4e701f20f53836c75b78740cabc3e29706f3ad17bb3ad377069c9b305edd4113" exitCode=0 Dec 03 13:50:50 crc kubenswrapper[4677]: I1203 13:50:50.346452 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4sxv" event={"ID":"04d74a24-b7f5-4026-bea7-7eb2caef99ba","Type":"ContainerDied","Data":"4e701f20f53836c75b78740cabc3e29706f3ad17bb3ad377069c9b305edd4113"} Dec 03 13:50:52 crc kubenswrapper[4677]: I1203 13:50:52.358067 4677 generic.go:334] "Generic (PLEG): container finished" podID="33e127ec-407a-4911-9478-b54509612df8" containerID="11443ad15a7bd1a62b97fe0807f0d22985824db963597d82132adb8f0d375113" exitCode=0 Dec 03 13:50:52 crc kubenswrapper[4677]: I1203 13:50:52.358147 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhghx" event={"ID":"33e127ec-407a-4911-9478-b54509612df8","Type":"ContainerDied","Data":"11443ad15a7bd1a62b97fe0807f0d22985824db963597d82132adb8f0d375113"} Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.659758 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.804801 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-utilities\") pod \"33e127ec-407a-4911-9478-b54509612df8\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.804894 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-catalog-content\") pod \"33e127ec-407a-4911-9478-b54509612df8\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.804934 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stfmk\" (UniqueName: \"kubernetes.io/projected/33e127ec-407a-4911-9478-b54509612df8-kube-api-access-stfmk\") pod \"33e127ec-407a-4911-9478-b54509612df8\" (UID: \"33e127ec-407a-4911-9478-b54509612df8\") " Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.807015 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-utilities" (OuterVolumeSpecName: "utilities") pod "33e127ec-407a-4911-9478-b54509612df8" (UID: "33e127ec-407a-4911-9478-b54509612df8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.812579 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e127ec-407a-4911-9478-b54509612df8-kube-api-access-stfmk" (OuterVolumeSpecName: "kube-api-access-stfmk") pod "33e127ec-407a-4911-9478-b54509612df8" (UID: "33e127ec-407a-4911-9478-b54509612df8"). InnerVolumeSpecName "kube-api-access-stfmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.854592 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33e127ec-407a-4911-9478-b54509612df8" (UID: "33e127ec-407a-4911-9478-b54509612df8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.906182 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.906230 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stfmk\" (UniqueName: \"kubernetes.io/projected/33e127ec-407a-4911-9478-b54509612df8-kube-api-access-stfmk\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:56 crc kubenswrapper[4677]: I1203 13:50:56.906244 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e127ec-407a-4911-9478-b54509612df8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:57 crc kubenswrapper[4677]: I1203 13:50:57.396029 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhghx" event={"ID":"33e127ec-407a-4911-9478-b54509612df8","Type":"ContainerDied","Data":"e1bd24574cd3ffba918b7b68a5c63e459d4bf8a16cf1ad0d4aa59b50e6fdf978"} Dec 03 13:50:57 crc kubenswrapper[4677]: I1203 13:50:57.396079 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhghx" Dec 03 13:50:57 crc kubenswrapper[4677]: I1203 13:50:57.396375 4677 scope.go:117] "RemoveContainer" containerID="11443ad15a7bd1a62b97fe0807f0d22985824db963597d82132adb8f0d375113" Dec 03 13:50:57 crc kubenswrapper[4677]: I1203 13:50:57.424476 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vhghx"] Dec 03 13:50:57 crc kubenswrapper[4677]: I1203 13:50:57.428639 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vhghx"] Dec 03 13:50:57 crc kubenswrapper[4677]: I1203 13:50:57.981761 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33e127ec-407a-4911-9478-b54509612df8" path="/var/lib/kubelet/pods/33e127ec-407a-4911-9478-b54509612df8/volumes" Dec 03 13:51:00 crc kubenswrapper[4677]: I1203 13:51:00.493369 4677 scope.go:117] "RemoveContainer" containerID="07c2c6bbe4e24779b55d60ba31e5e01aa12fa36e6be1c2748ca1a73527284617" Dec 03 13:51:02 crc kubenswrapper[4677]: I1203 13:51:02.692061 4677 scope.go:117] "RemoveContainer" containerID="57e4287dc21ec268ee9bd6a3782bdb166b8c9099b8bd585b23347da72a30e1d8" Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.478692 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4sxv" event={"ID":"04d74a24-b7f5-4026-bea7-7eb2caef99ba","Type":"ContainerStarted","Data":"ad0e62d1ac27c30b090514864fb494bdfcd7b343abb85ec53f276898d3597532"} Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.481834 4677 generic.go:334] "Generic (PLEG): container finished" podID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerID="c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b" exitCode=0 Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.481876 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbnmw" event={"ID":"96689c77-5d8c-46c9-a87b-fc98bf5797ed","Type":"ContainerDied","Data":"c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b"} Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.484207 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhnw" event={"ID":"d9a210ab-e032-4834-bb1c-b9835456d960","Type":"ContainerStarted","Data":"1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a"} Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.488068 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th96p" event={"ID":"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87","Type":"ContainerStarted","Data":"f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a"} Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.496728 4677 generic.go:334] "Generic (PLEG): container finished" podID="73e747c5-106b-479c-9264-558bf7783d87" containerID="84e53a5a21cd43fe8f6a743c206621b0e9765a1d7ea2bc55d96d3327b59ebf08" exitCode=0 Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.496762 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvzwn" event={"ID":"73e747c5-106b-479c-9264-558bf7783d87","Type":"ContainerDied","Data":"84e53a5a21cd43fe8f6a743c206621b0e9765a1d7ea2bc55d96d3327b59ebf08"} Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.525395 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w4sxv" podStartSLOduration=3.286814463 podStartE2EDuration="1m35.525381849s" podCreationTimestamp="2025-12-03 13:49:28 +0000 UTC" firstStartedPulling="2025-12-03 13:49:30.475327611 +0000 UTC m=+161.221660056" lastFinishedPulling="2025-12-03 13:51:02.713894987 +0000 UTC m=+253.460227442" observedRunningTime="2025-12-03 13:51:03.500873105 +0000 UTC m=+254.247205570" watchObservedRunningTime="2025-12-03 13:51:03.525381849 +0000 UTC m=+254.271714304" Dec 03 13:51:03 crc kubenswrapper[4677]: I1203 13:51:03.596670 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-th96p" podStartSLOduration=2.201534505 podStartE2EDuration="1m34.596651829s" podCreationTimestamp="2025-12-03 13:49:29 +0000 UTC" firstStartedPulling="2025-12-03 13:49:30.483025368 +0000 UTC m=+161.229357823" lastFinishedPulling="2025-12-03 13:51:02.878142692 +0000 UTC m=+253.624475147" observedRunningTime="2025-12-03 13:51:03.566927119 +0000 UTC m=+254.313259594" watchObservedRunningTime="2025-12-03 13:51:03.596651829 +0000 UTC m=+254.342984284" Dec 03 13:51:04 crc kubenswrapper[4677]: I1203 13:51:04.505095 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbnmw" event={"ID":"96689c77-5d8c-46c9-a87b-fc98bf5797ed","Type":"ContainerStarted","Data":"1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0"} Dec 03 13:51:04 crc kubenswrapper[4677]: I1203 13:51:04.508623 4677 generic.go:334] "Generic (PLEG): container finished" podID="d9a210ab-e032-4834-bb1c-b9835456d960" containerID="1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a" exitCode=0 Dec 03 13:51:04 crc kubenswrapper[4677]: I1203 13:51:04.508715 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhnw" event={"ID":"d9a210ab-e032-4834-bb1c-b9835456d960","Type":"ContainerDied","Data":"1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a"} Dec 03 13:51:04 crc kubenswrapper[4677]: I1203 13:51:04.510994 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvzwn" event={"ID":"73e747c5-106b-479c-9264-558bf7783d87","Type":"ContainerStarted","Data":"b0b995bfb18ac4a67f4d65b15cce1ff133818f8fb25278d8d04d2b0cf1c801de"} Dec 03 13:51:04 crc kubenswrapper[4677]: I1203 13:51:04.531688 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dbnmw" podStartSLOduration=3.8179673899999997 podStartE2EDuration="1m38.531665443s" podCreationTimestamp="2025-12-03 13:49:26 +0000 UTC" firstStartedPulling="2025-12-03 13:49:29.458236151 +0000 UTC m=+160.204568616" lastFinishedPulling="2025-12-03 13:51:04.171934204 +0000 UTC m=+254.918266669" observedRunningTime="2025-12-03 13:51:04.527214781 +0000 UTC m=+255.273547256" watchObservedRunningTime="2025-12-03 13:51:04.531665443 +0000 UTC m=+255.277997918" Dec 03 13:51:04 crc kubenswrapper[4677]: I1203 13:51:04.545829 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pvzwn" podStartSLOduration=4.014788975 podStartE2EDuration="1m38.545814145s" podCreationTimestamp="2025-12-03 13:49:26 +0000 UTC" firstStartedPulling="2025-12-03 13:49:29.435002696 +0000 UTC m=+160.181335151" lastFinishedPulling="2025-12-03 13:51:03.966027866 +0000 UTC m=+254.712360321" observedRunningTime="2025-12-03 13:51:04.542802829 +0000 UTC m=+255.289135294" watchObservedRunningTime="2025-12-03 13:51:04.545814145 +0000 UTC m=+255.292146600" Dec 03 13:51:05 crc kubenswrapper[4677]: I1203 13:51:05.518413 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhnw" event={"ID":"d9a210ab-e032-4834-bb1c-b9835456d960","Type":"ContainerStarted","Data":"b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db"} Dec 03 13:51:05 crc kubenswrapper[4677]: I1203 13:51:05.535194 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2qhnw" podStartSLOduration=3.051646175 podStartE2EDuration="1m36.535178737s" podCreationTimestamp="2025-12-03 13:49:29 +0000 UTC" firstStartedPulling="2025-12-03 13:49:31.537886719 +0000 UTC m=+162.284219174" lastFinishedPulling="2025-12-03 13:51:05.021419281 +0000 UTC m=+255.767751736" observedRunningTime="2025-12-03 13:51:05.533974969 +0000 UTC m=+256.280307444" watchObservedRunningTime="2025-12-03 13:51:05.535178737 +0000 UTC m=+256.281511192" Dec 03 13:51:06 crc kubenswrapper[4677]: I1203 13:51:06.796784 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:51:06 crc kubenswrapper[4677]: I1203 13:51:06.797834 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:51:06 crc kubenswrapper[4677]: I1203 13:51:06.836630 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:51:07 crc kubenswrapper[4677]: I1203 13:51:07.221441 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:51:07 crc kubenswrapper[4677]: I1203 13:51:07.221495 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:51:07 crc kubenswrapper[4677]: I1203 13:51:07.265792 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:51:07 crc kubenswrapper[4677]: I1203 13:51:07.674541 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntqz8"] Dec 03 13:51:09 crc kubenswrapper[4677]: I1203 13:51:09.063501 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:51:09 crc kubenswrapper[4677]: I1203 13:51:09.063597 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:51:09 crc kubenswrapper[4677]: I1203 13:51:09.100773 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:51:09 crc kubenswrapper[4677]: I1203 13:51:09.588280 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:51:09 crc kubenswrapper[4677]: I1203 13:51:09.899055 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:51:09 crc kubenswrapper[4677]: I1203 13:51:09.899108 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:51:09 crc kubenswrapper[4677]: I1203 13:51:09.947152 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:51:10 crc kubenswrapper[4677]: I1203 13:51:10.200148 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:51:10 crc kubenswrapper[4677]: I1203 13:51:10.200284 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:51:10 crc kubenswrapper[4677]: I1203 13:51:10.599562 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.119523 4677 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.120135 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="registry-server" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120152 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="registry-server" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.120165 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="extract-utilities" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120173 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="extract-utilities" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.120200 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f965bc9-867b-4abe-aca7-8a39f1863311" containerName="pruner" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120208 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f965bc9-867b-4abe-aca7-8a39f1863311" containerName="pruner" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.120224 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="extract-content" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120231 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="extract-content" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120353 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f965bc9-867b-4abe-aca7-8a39f1863311" containerName="pruner" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120369 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e127ec-407a-4911-9478-b54509612df8" containerName="registry-server" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120817 4677 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.120915 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.121140 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997" gracePeriod=15 Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.121161 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b" gracePeriod=15 Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.121199 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3" gracePeriod=15 Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.121212 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c" gracePeriod=15 Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.121223 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04" gracePeriod=15 Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.121508 4677 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.122700 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.122717 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.122727 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.122734 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.122747 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.122753 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.122771 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.122779 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.122786 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.122792 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.122800 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.122807 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.122933 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.123031 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.123044 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.123054 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.123229 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.123244 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.123400 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.123409 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.204042 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.204088 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.204142 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.204177 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.204206 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.242515 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2qhnw" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="registry-server" probeResult="failure" output=< Dec 03 13:51:11 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 13:51:11 crc kubenswrapper[4677]: > Dec 03 13:51:11 crc kubenswrapper[4677]: E1203 13:51:11.243227 4677 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.162:6443: connect: connection refused" event=< Dec 03 13:51:11 crc kubenswrapper[4677]: &Event{ObjectMeta:{redhat-operators-2qhnw.187db8d905b98c5b openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-2qhnw,UID:d9a210ab-e032-4834-bb1c-b9835456d960,APIVersion:v1,ResourceVersion:28645,FieldPath:spec.containers{registry-server},},Reason:Unhealthy,Message:Startup probe failed: timeout: failed to connect service ":50051" within 1s Dec 03 13:51:11 crc kubenswrapper[4677]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:51:11.242562651 +0000 UTC m=+261.988895106,LastTimestamp:2025-12-03 13:51:11.242562651 +0000 UTC m=+261.988895106,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 03 13:51:11 crc kubenswrapper[4677]: > Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305401 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305460 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305510 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305532 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305536 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305595 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305548 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305692 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305752 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305771 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305790 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305808 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.305829 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.407338 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.407388 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.407414 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.407441 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.407496 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:11 crc kubenswrapper[4677]: I1203 13:51:11.407501 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.565193 4677 generic.go:334] "Generic (PLEG): container finished" podID="536f457b-8005-409a-84e8-967f9d6f968d" containerID="e29148d7b9ff52704c1a5ef775f2af31ab5c35747bea6896dcf1c8d075180ab1" exitCode=0 Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.565297 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"536f457b-8005-409a-84e8-967f9d6f968d","Type":"ContainerDied","Data":"e29148d7b9ff52704c1a5ef775f2af31ab5c35747bea6896dcf1c8d075180ab1"} Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.566890 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.568979 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.570363 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.571760 4677 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b" exitCode=0 Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.571788 4677 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04" exitCode=0 Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.571798 4677 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c" exitCode=0 Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.571809 4677 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3" exitCode=2 Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.571819 4677 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997" exitCode=0 Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.571857 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd295efabb1aced7c89f9810c78d667655ab213d76df3f06ebf456e96bcef6d1" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.571873 4677 scope.go:117] "RemoveContainer" containerID="888bf2bef4d615a6d0dda2c10b0279d4b5da944c31308bb947c544ab3a1c732f" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.599592 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.600427 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.601042 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.601408 4677 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.743634 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.743826 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.744244 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.744368 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.744557 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.744655 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.745067 4677 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.745162 4677 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.745224 4677 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:13 crc kubenswrapper[4677]: I1203 13:51:13.983918 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.579574 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.581792 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.583145 4677 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.583589 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.584818 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.585272 4677 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.809038 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.809761 4677 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.810049 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.960219 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-kubelet-dir\") pod \"536f457b-8005-409a-84e8-967f9d6f968d\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.960287 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/536f457b-8005-409a-84e8-967f9d6f968d-kube-api-access\") pod \"536f457b-8005-409a-84e8-967f9d6f968d\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.960318 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-var-lock\") pod \"536f457b-8005-409a-84e8-967f9d6f968d\" (UID: \"536f457b-8005-409a-84e8-967f9d6f968d\") " Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.960575 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-var-lock" (OuterVolumeSpecName: "var-lock") pod "536f457b-8005-409a-84e8-967f9d6f968d" (UID: "536f457b-8005-409a-84e8-967f9d6f968d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.960604 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "536f457b-8005-409a-84e8-967f9d6f968d" (UID: "536f457b-8005-409a-84e8-967f9d6f968d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:51:14 crc kubenswrapper[4677]: I1203 13:51:14.965584 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536f457b-8005-409a-84e8-967f9d6f968d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "536f457b-8005-409a-84e8-967f9d6f968d" (UID: "536f457b-8005-409a-84e8-967f9d6f968d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.061668 4677 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.061721 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/536f457b-8005-409a-84e8-967f9d6f968d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.061741 4677 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/536f457b-8005-409a-84e8-967f9d6f968d-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.587853 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"536f457b-8005-409a-84e8-967f9d6f968d","Type":"ContainerDied","Data":"dbdf09959deaeaa84548ac87826dc8634bbcc397b450c1ae310025e9234ab1b0"} Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.587895 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbdf09959deaeaa84548ac87826dc8634bbcc397b450c1ae310025e9234ab1b0" Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.587993 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.600282 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:15 crc kubenswrapper[4677]: I1203 13:51:15.600495 4677 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:16 crc kubenswrapper[4677]: E1203 13:51:16.150080 4677 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:16 crc kubenswrapper[4677]: I1203 13:51:16.151463 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:16 crc kubenswrapper[4677]: W1203 13:51:16.179107 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-4c959ec88b781eb64bd9c7a633fcdda9fcdd3e6760a3e527f050564b6af966a3 WatchSource:0}: Error finding container 4c959ec88b781eb64bd9c7a633fcdda9fcdd3e6760a3e527f050564b6af966a3: Status 404 returned error can't find the container with id 4c959ec88b781eb64bd9c7a633fcdda9fcdd3e6760a3e527f050564b6af966a3 Dec 03 13:51:16 crc kubenswrapper[4677]: I1203 13:51:16.595460 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4c959ec88b781eb64bd9c7a633fcdda9fcdd3e6760a3e527f050564b6af966a3"} Dec 03 13:51:16 crc kubenswrapper[4677]: I1203 13:51:16.852040 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:51:16 crc kubenswrapper[4677]: I1203 13:51:16.852803 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:16 crc kubenswrapper[4677]: I1203 13:51:16.853049 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: I1203 13:51:17.271350 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:51:17 crc kubenswrapper[4677]: I1203 13:51:17.273179 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: I1203 13:51:17.273606 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: I1203 13:51:17.273901 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: E1203 13:51:17.850935 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:51:17Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:51:17Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:51:17Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T13:51:17Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: E1203 13:51:17.851148 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: E1203 13:51:17.851286 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: E1203 13:51:17.851418 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: E1203 13:51:17.851545 4677 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:17 crc kubenswrapper[4677]: E1203 13:51:17.851556 4677 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 13:51:18 crc kubenswrapper[4677]: E1203 13:51:18.796404 4677 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.162:6443: connect: connection refused" event=< Dec 03 13:51:18 crc kubenswrapper[4677]: &Event{ObjectMeta:{redhat-operators-2qhnw.187db8d905b98c5b openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-2qhnw,UID:d9a210ab-e032-4834-bb1c-b9835456d960,APIVersion:v1,ResourceVersion:28645,FieldPath:spec.containers{registry-server},},Reason:Unhealthy,Message:Startup probe failed: timeout: failed to connect service ":50051" within 1s Dec 03 13:51:18 crc kubenswrapper[4677]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 13:51:11.242562651 +0000 UTC m=+261.988895106,LastTimestamp:2025-12-03 13:51:11.242562651 +0000 UTC m=+261.988895106,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 03 13:51:18 crc kubenswrapper[4677]: > Dec 03 13:51:19 crc kubenswrapper[4677]: I1203 13:51:19.979404 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:19 crc kubenswrapper[4677]: I1203 13:51:19.980604 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:19 crc kubenswrapper[4677]: I1203 13:51:19.981175 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.241874 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.243020 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.243560 4677 status_manager.go:851] "Failed to get status for pod" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" pod="openshift-marketplace/redhat-operators-2qhnw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2qhnw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.243879 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.244162 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.281202 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.281709 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.282072 4677 status_manager.go:851] "Failed to get status for pod" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" pod="openshift-marketplace/redhat-operators-2qhnw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2qhnw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.282305 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.282462 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: E1203 13:51:20.900180 4677 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: E1203 13:51:20.900617 4677 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: E1203 13:51:20.901016 4677 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: E1203 13:51:20.901330 4677 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: E1203 13:51:20.901612 4677 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:20 crc kubenswrapper[4677]: I1203 13:51:20.901651 4677 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 13:51:20 crc kubenswrapper[4677]: E1203 13:51:20.901984 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="200ms" Dec 03 13:51:21 crc kubenswrapper[4677]: E1203 13:51:21.103181 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="400ms" Dec 03 13:51:21 crc kubenswrapper[4677]: E1203 13:51:21.504157 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="800ms" Dec 03 13:51:22 crc kubenswrapper[4677]: E1203 13:51:22.304893 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="1.6s" Dec 03 13:51:23 crc kubenswrapper[4677]: E1203 13:51:23.906756 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="3.2s" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.975541 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.976681 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.977203 4677 status_manager.go:851] "Failed to get status for pod" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" pod="openshift-marketplace/redhat-operators-2qhnw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2qhnw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.977513 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.977886 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.990724 4677 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.990754 4677 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:24 crc kubenswrapper[4677]: E1203 13:51:24.991195 4677 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:24 crc kubenswrapper[4677]: I1203 13:51:24.991736 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:25 crc kubenswrapper[4677]: I1203 13:51:25.656028 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0da5c19a2eac860e4ffea5b064d03986bdb89f221f46f135b98f521d990bdc49"} Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.448573 4677 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.448656 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.662578 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"05d1717a2d96d601d9e6c550b062961f67687b40ffc13b5b2cfbace14490920c"} Dec 03 13:51:26 crc kubenswrapper[4677]: E1203 13:51:26.663885 4677 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.663920 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.664320 4677 status_manager.go:851] "Failed to get status for pod" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" pod="openshift-marketplace/redhat-operators-2qhnw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2qhnw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.665243 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.665432 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"64b55127bc601e41d3d9c7d8622858f69486ea313b992b632d16daca71040784"} Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.665838 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.670771 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.670812 4677 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035" exitCode=1 Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.670837 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035"} Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.671387 4677 scope.go:117] "RemoveContainer" containerID="42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.671581 4677 status_manager.go:851] "Failed to get status for pod" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" pod="openshift-marketplace/redhat-operators-2qhnw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2qhnw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.672101 4677 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.672478 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.672725 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:26 crc kubenswrapper[4677]: I1203 13:51:26.672869 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: E1203 13:51:27.108095 4677 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="6.4s" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.676696 4677 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="64b55127bc601e41d3d9c7d8622858f69486ea313b992b632d16daca71040784" exitCode=0 Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.676796 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"64b55127bc601e41d3d9c7d8622858f69486ea313b992b632d16daca71040784"} Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.677014 4677 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.677034 4677 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.677426 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: E1203 13:51:27.677460 4677 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.677612 4677 status_manager.go:851] "Failed to get status for pod" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" pod="openshift-marketplace/redhat-operators-2qhnw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2qhnw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.677882 4677 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.678298 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.678858 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.680248 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.680317 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f5071c3757b4cdb8e613f663ffd568ada887a6a40fa2ead131e85042a780db91"} Dec 03 13:51:27 crc kubenswrapper[4677]: E1203 13:51:27.680892 4677 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.681127 4677 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.681480 4677 status_manager.go:851] "Failed to get status for pod" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" pod="openshift-marketplace/certified-operators-dbnmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dbnmw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.681975 4677 status_manager.go:851] "Failed to get status for pod" podUID="536f457b-8005-409a-84e8-967f9d6f968d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.682421 4677 status_manager.go:851] "Failed to get status for pod" podUID="73e747c5-106b-479c-9264-558bf7783d87" pod="openshift-marketplace/certified-operators-pvzwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-pvzwn\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:27 crc kubenswrapper[4677]: I1203 13:51:27.682934 4677 status_manager.go:851] "Failed to get status for pod" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" pod="openshift-marketplace/redhat-operators-2qhnw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2qhnw\": dial tcp 38.102.83.162:6443: connect: connection refused" Dec 03 13:51:28 crc kubenswrapper[4677]: I1203 13:51:28.688008 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3ce3b850c542950a24644e765670a6afefebd0da9835fe822d955d32112af9db"} Dec 03 13:51:31 crc kubenswrapper[4677]: I1203 13:51:29.696902 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dfd2ec105b69d47b1d38c596e609f63c00d1fef8a711ce7f50c6dd3e35ec22cc"} Dec 03 13:51:31 crc kubenswrapper[4677]: I1203 13:51:29.713372 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:51:31 crc kubenswrapper[4677]: I1203 13:51:29.717533 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:51:31 crc kubenswrapper[4677]: I1203 13:51:30.702120 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:51:31 crc kubenswrapper[4677]: I1203 13:51:31.712538 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f2ae7f838baa7e8618f606a35b808777f745ad04c521b0849631597a744c1295"} Dec 03 13:51:31 crc kubenswrapper[4677]: I1203 13:51:31.713151 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"848daf8c63ec534a594c5af5ce27dd56b46d26bc89e2782432d27a1a319a2d13"} Dec 03 13:51:32 crc kubenswrapper[4677]: I1203 13:51:32.701543 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" podUID="c2c20cb8-d990-4756-ad31-56245b583279" containerName="oauth-openshift" containerID="cri-o://e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b" gracePeriod=15 Dec 03 13:51:32 crc kubenswrapper[4677]: I1203 13:51:32.728110 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2f61f515570f1f0e8388c77af44d9bc0817222b28e1dd7bf2f72f00c5460ee72"} Dec 03 13:51:32 crc kubenswrapper[4677]: I1203 13:51:32.728308 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:32 crc kubenswrapper[4677]: I1203 13:51:32.728405 4677 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:32 crc kubenswrapper[4677]: I1203 13:51:32.728429 4677 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:32 crc kubenswrapper[4677]: I1203 13:51:32.738308 4677 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.106643 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.156888 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsj5r\" (UniqueName: \"kubernetes.io/projected/c2c20cb8-d990-4756-ad31-56245b583279-kube-api-access-hsj5r\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157025 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-provider-selection\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157082 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-cliconfig\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157122 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2c20cb8-d990-4756-ad31-56245b583279-audit-dir\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157167 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-error\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157216 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-router-certs\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157301 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-audit-policies\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157335 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2c20cb8-d990-4756-ad31-56245b583279-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157346 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-idp-0-file-data\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157451 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-ocp-branding-template\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157507 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-session\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157550 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-service-ca\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157577 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-trusted-ca-bundle\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157603 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-serving-cert\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.157627 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-login\") pod \"c2c20cb8-d990-4756-ad31-56245b583279\" (UID: \"c2c20cb8-d990-4756-ad31-56245b583279\") " Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.158057 4677 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2c20cb8-d990-4756-ad31-56245b583279-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.159102 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.159917 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.160686 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.160938 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.164690 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.165021 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.165900 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.166544 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.167127 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.167458 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.167905 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.168619 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c20cb8-d990-4756-ad31-56245b583279-kube-api-access-hsj5r" (OuterVolumeSpecName: "kube-api-access-hsj5r") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "kube-api-access-hsj5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.169425 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c2c20cb8-d990-4756-ad31-56245b583279" (UID: "c2c20cb8-d990-4756-ad31-56245b583279"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259395 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259430 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259460 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259470 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259480 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259489 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsj5r\" (UniqueName: \"kubernetes.io/projected/c2c20cb8-d990-4756-ad31-56245b583279-kube-api-access-hsj5r\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259499 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259509 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259518 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259526 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259537 4677 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2c20cb8-d990-4756-ad31-56245b583279-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259546 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.259556 4677 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2c20cb8-d990-4756-ad31-56245b583279-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.734548 4677 generic.go:334] "Generic (PLEG): container finished" podID="c2c20cb8-d990-4756-ad31-56245b583279" containerID="e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b" exitCode=0 Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.734619 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" event={"ID":"c2c20cb8-d990-4756-ad31-56245b583279","Type":"ContainerDied","Data":"e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b"} Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.734690 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.735905 4677 scope.go:117] "RemoveContainer" containerID="e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.735795 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntqz8" event={"ID":"c2c20cb8-d990-4756-ad31-56245b583279","Type":"ContainerDied","Data":"ea200524ede1c3f37a843e83822a3c91bb4c3d8818526f955b2347845365d270"} Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.737312 4677 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.737470 4677 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.764152 4677 scope.go:117] "RemoveContainer" containerID="e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b" Dec 03 13:51:33 crc kubenswrapper[4677]: E1203 13:51:33.765101 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b\": container with ID starting with e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b not found: ID does not exist" containerID="e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b" Dec 03 13:51:33 crc kubenswrapper[4677]: I1203 13:51:33.765149 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b"} err="failed to get container status \"e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b\": rpc error: code = NotFound desc = could not find container \"e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b\": container with ID starting with e2c99431e2ade33df4abb4060f4afe0dcf4324e8aec635a0c937b8a2c3aefb0b not found: ID does not exist" Dec 03 13:51:34 crc kubenswrapper[4677]: I1203 13:51:34.975840 4677 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="deccfaa3-2f6d-4c70-88fd-fce515af3819" Dec 03 13:51:35 crc kubenswrapper[4677]: E1203 13:51:35.374739 4677 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 03 13:51:35 crc kubenswrapper[4677]: E1203 13:51:35.622108 4677 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 03 13:51:35 crc kubenswrapper[4677]: E1203 13:51:35.787149 4677 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 03 13:51:40 crc kubenswrapper[4677]: I1203 13:51:40.900164 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 13:51:45 crc kubenswrapper[4677]: I1203 13:51:45.305943 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 13:51:45 crc kubenswrapper[4677]: I1203 13:51:45.754838 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 13:51:45 crc kubenswrapper[4677]: I1203 13:51:45.990091 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 13:51:46 crc kubenswrapper[4677]: I1203 13:51:46.017592 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 13:51:46 crc kubenswrapper[4677]: I1203 13:51:46.172779 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 13:51:46 crc kubenswrapper[4677]: I1203 13:51:46.482024 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 13:51:46 crc kubenswrapper[4677]: I1203 13:51:46.705020 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.055039 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.080928 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.172059 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.305030 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.384729 4677 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.407989 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.536656 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.591836 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.595365 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.661649 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.810278 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.812587 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.818427 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 13:51:47 crc kubenswrapper[4677]: I1203 13:51:47.856101 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.000722 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.174166 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.204893 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.228493 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.253761 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.359002 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.424610 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.619533 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.663185 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.713868 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.730238 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.732035 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.815607 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.920489 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 13:51:48 crc kubenswrapper[4677]: I1203 13:51:48.931266 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.027054 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.082362 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.266911 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.289549 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.329901 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.394903 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.449663 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.480480 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.500542 4677 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.506404 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntqz8","openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.506482 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.506843 4677 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.506876 4677 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="88e2b75b-db00-4a56-8e0b-0dc9abb880d2" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.512213 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.514535 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.524500 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.524480877 podStartE2EDuration="17.524480877s" podCreationTimestamp="2025-12-03 13:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:51:49.52328618 +0000 UTC m=+300.269618645" watchObservedRunningTime="2025-12-03 13:51:49.524480877 +0000 UTC m=+300.270813342" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.558244 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.622766 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.653532 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.704508 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.764976 4677 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.766386 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.802356 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.904461 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.983904 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c20cb8-d990-4756-ad31-56245b583279" path="/var/lib/kubelet/pods/c2c20cb8-d990-4756-ad31-56245b583279/volumes" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.992526 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.992568 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.997091 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 13:51:49 crc kubenswrapper[4677]: I1203 13:51:49.997513 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.007209 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.113631 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.116032 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.281857 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.292402 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.337945 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.381830 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.411690 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.549798 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.651404 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.738165 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.792113 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.853114 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.927085 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.986868 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 13:51:50 crc kubenswrapper[4677]: I1203 13:51:50.990045 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.032924 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.095222 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.137680 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.186474 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.233246 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.245668 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.287660 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.317749 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.331746 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.349436 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.371815 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.393408 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.394788 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.502568 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.505592 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.509461 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.682099 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.695250 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.701344 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.723067 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.790763 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.822494 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.876719 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.921414 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.963530 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 13:51:51 crc kubenswrapper[4677]: I1203 13:51:51.991424 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.201586 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.252881 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.344487 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.531465 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.594847 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.632100 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.671720 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.699296 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.846370 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.851627 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 13:51:52 crc kubenswrapper[4677]: I1203 13:51:52.981567 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.077649 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.155930 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.182083 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.226622 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.250596 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-9448d8947-cdz7q"] Dec 03 13:51:53 crc kubenswrapper[4677]: E1203 13:51:53.251057 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c20cb8-d990-4756-ad31-56245b583279" containerName="oauth-openshift" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.251092 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c20cb8-d990-4756-ad31-56245b583279" containerName="oauth-openshift" Dec 03 13:51:53 crc kubenswrapper[4677]: E1203 13:51:53.251111 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536f457b-8005-409a-84e8-967f9d6f968d" containerName="installer" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.251137 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="536f457b-8005-409a-84e8-967f9d6f968d" containerName="installer" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.251256 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="536f457b-8005-409a-84e8-967f9d6f968d" containerName="installer" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.251273 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c20cb8-d990-4756-ad31-56245b583279" containerName="oauth-openshift" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.251912 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.254039 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.254433 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.254905 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.255028 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.255575 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.255622 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.255893 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.256999 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.257316 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.257098 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.257893 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.260924 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.280865 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.282380 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.285345 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.296882 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.310104 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.320810 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.321253 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-router-certs\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.321416 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ae79228-b546-472d-9a25-9161ecc2f296-audit-dir\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.321568 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-service-ca\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.321702 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.321852 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc6lh\" (UniqueName: \"kubernetes.io/projected/7ae79228-b546-472d-9a25-9161ecc2f296-kube-api-access-bc6lh\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.321995 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.322177 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.322288 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-error\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.322379 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-login\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.322470 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.322574 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-session\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.322712 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-audit-policies\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.322815 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.365260 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.382703 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.414270 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.414310 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.421849 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424084 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-router-certs\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424147 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ae79228-b546-472d-9a25-9161ecc2f296-audit-dir\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424184 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-service-ca\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424207 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424248 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc6lh\" (UniqueName: \"kubernetes.io/projected/7ae79228-b546-472d-9a25-9161ecc2f296-kube-api-access-bc6lh\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424279 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424298 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ae79228-b546-472d-9a25-9161ecc2f296-audit-dir\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.424306 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425200 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-service-ca\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425232 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-error\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425306 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-login\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425350 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425428 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-session\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425506 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-audit-policies\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425551 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.425691 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.426341 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-audit-policies\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.426597 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.426910 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.430459 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.431037 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-router-certs\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.431943 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-login\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.432406 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-error\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.433434 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-session\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.434714 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.435326 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.435559 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ae79228-b546-472d-9a25-9161ecc2f296-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.444753 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc6lh\" (UniqueName: \"kubernetes.io/projected/7ae79228-b546-472d-9a25-9161ecc2f296-kube-api-access-bc6lh\") pod \"oauth-openshift-9448d8947-cdz7q\" (UID: \"7ae79228-b546-472d-9a25-9161ecc2f296\") " pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.556380 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.593143 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.650256 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.727526 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.777366 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.854700 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.855536 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.889897 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 13:51:53 crc kubenswrapper[4677]: I1203 13:51:53.922818 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.025244 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.085301 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.101458 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.143178 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.205202 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.489431 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.490653 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.511070 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.548978 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.561892 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.624919 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.650158 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.670147 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.741904 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.787797 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 13:51:54 crc kubenswrapper[4677]: I1203 13:51:54.795463 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.111600 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.122704 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.171478 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9448d8947-cdz7q"] Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.200697 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.350669 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.397278 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.449148 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.499547 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.536383 4677 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.545578 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.558462 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.561509 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.627194 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9448d8947-cdz7q"] Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.721426 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.723027 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.859224 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 13:51:55 crc kubenswrapper[4677]: I1203 13:51:55.876029 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" event={"ID":"7ae79228-b546-472d-9a25-9161ecc2f296","Type":"ContainerStarted","Data":"94e4930f7cc348467f8f3120ca974db8c5c476e75b0581fdbdb42eff353f0282"} Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.210353 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.258156 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.494400 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.498934 4677 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.499176 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://05d1717a2d96d601d9e6c550b062961f67687b40ffc13b5b2cfbace14490920c" gracePeriod=5 Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.547265 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.641581 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.648347 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.681856 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.729784 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.752918 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.781173 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.801960 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.825368 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.833976 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.885612 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" event={"ID":"7ae79228-b546-472d-9a25-9161ecc2f296","Type":"ContainerStarted","Data":"3d16b970fbeb8664312694f8054d357c40e23240dae7cf6f72edd6633c4316ad"} Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.886371 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:56 crc kubenswrapper[4677]: I1203 13:51:56.914232 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" podStartSLOduration=49.914211467 podStartE2EDuration="49.914211467s" podCreationTimestamp="2025-12-03 13:51:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:51:56.907568377 +0000 UTC m=+307.653900842" watchObservedRunningTime="2025-12-03 13:51:56.914211467 +0000 UTC m=+307.660543922" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.000057 4677 patch_prober.go:28] interesting pod/oauth-openshift-9448d8947-cdz7q container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": read tcp 10.217.0.2:35932->10.217.0.56:6443: read: connection reset by peer" start-of-body= Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.000132 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" podUID="7ae79228-b546-472d-9a25-9161ecc2f296" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": read tcp 10.217.0.2:35932->10.217.0.56:6443: read: connection reset by peer" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.022312 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.076773 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.119910 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.254360 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.264766 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.272150 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.277280 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.280021 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.477321 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.571301 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.583136 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.601644 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.659599 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.730820 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.863420 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.896980 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-9448d8947-cdz7q_7ae79228-b546-472d-9a25-9161ecc2f296/oauth-openshift/0.log" Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.897054 4677 generic.go:334] "Generic (PLEG): container finished" podID="7ae79228-b546-472d-9a25-9161ecc2f296" containerID="3d16b970fbeb8664312694f8054d357c40e23240dae7cf6f72edd6633c4316ad" exitCode=255 Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.897087 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" event={"ID":"7ae79228-b546-472d-9a25-9161ecc2f296","Type":"ContainerDied","Data":"3d16b970fbeb8664312694f8054d357c40e23240dae7cf6f72edd6633c4316ad"} Dec 03 13:51:57 crc kubenswrapper[4677]: I1203 13:51:57.897526 4677 scope.go:117] "RemoveContainer" containerID="3d16b970fbeb8664312694f8054d357c40e23240dae7cf6f72edd6633c4316ad" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.078449 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.158584 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.206135 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.481605 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.630135 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.631107 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.750643 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.863693 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.904535 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-9448d8947-cdz7q_7ae79228-b546-472d-9a25-9161ecc2f296/oauth-openshift/0.log" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.904604 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" event={"ID":"7ae79228-b546-472d-9a25-9161ecc2f296","Type":"ContainerStarted","Data":"433689c32813b62319d06c8cf0b96b7a99064a45aecfb355ec40e53d41747540"} Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.904919 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.939003 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.946801 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.970616 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 13:51:58 crc kubenswrapper[4677]: I1203 13:51:58.986437 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.025451 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.048557 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.381686 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.448387 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.589574 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.614250 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.659066 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.962927 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 13:51:59 crc kubenswrapper[4677]: I1203 13:51:59.964275 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 13:52:00 crc kubenswrapper[4677]: I1203 13:52:00.091516 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 13:52:00 crc kubenswrapper[4677]: I1203 13:52:00.198203 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 13:52:00 crc kubenswrapper[4677]: I1203 13:52:00.204574 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 13:52:00 crc kubenswrapper[4677]: I1203 13:52:00.526386 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 13:52:00 crc kubenswrapper[4677]: I1203 13:52:00.666904 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 13:52:00 crc kubenswrapper[4677]: I1203 13:52:00.697482 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 13:52:01 crc kubenswrapper[4677]: I1203 13:52:01.922302 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 13:52:01 crc kubenswrapper[4677]: I1203 13:52:01.922590 4677 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="05d1717a2d96d601d9e6c550b062961f67687b40ffc13b5b2cfbace14490920c" exitCode=137 Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.393702 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.393774 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465027 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465342 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465483 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465510 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465648 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465687 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465714 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.465731 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.466021 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.466090 4677 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.466198 4677 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.466259 4677 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.474909 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.567991 4677 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.569409 4677 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.930628 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.930964 4677 scope.go:117] "RemoveContainer" containerID="05d1717a2d96d601d9e6c550b062961f67687b40ffc13b5b2cfbace14490920c" Dec 03 13:52:02 crc kubenswrapper[4677]: I1203 13:52:02.931104 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 13:52:03 crc kubenswrapper[4677]: I1203 13:52:03.983332 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 13:52:11 crc kubenswrapper[4677]: I1203 13:52:11.121586 4677 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 13:52:13 crc kubenswrapper[4677]: I1203 13:52:13.142189 4677 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 13:52:14 crc kubenswrapper[4677]: I1203 13:52:14.412316 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 13:52:14 crc kubenswrapper[4677]: I1203 13:52:14.509503 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 13:52:14 crc kubenswrapper[4677]: I1203 13:52:14.514372 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 13:52:21 crc kubenswrapper[4677]: I1203 13:52:21.495867 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 13:52:22 crc kubenswrapper[4677]: I1203 13:52:22.472654 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 13:52:22 crc kubenswrapper[4677]: I1203 13:52:22.521996 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 13:52:22 crc kubenswrapper[4677]: I1203 13:52:22.750742 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 13:52:23 crc kubenswrapper[4677]: I1203 13:52:23.116381 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 13:52:23 crc kubenswrapper[4677]: I1203 13:52:23.381586 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 13:52:23 crc kubenswrapper[4677]: I1203 13:52:23.994782 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 13:52:25 crc kubenswrapper[4677]: I1203 13:52:25.750163 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 13:52:26 crc kubenswrapper[4677]: I1203 13:52:26.279476 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 13:52:27 crc kubenswrapper[4677]: I1203 13:52:27.073657 4677 generic.go:334] "Generic (PLEG): container finished" podID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerID="9370c5745888934d7eaa524156e7b23861ec199211dc3d49521f028841ea3f51" exitCode=0 Dec 03 13:52:27 crc kubenswrapper[4677]: I1203 13:52:27.073738 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" event={"ID":"9dc01e2f-388f-41a2-a139-5d251ad6cda8","Type":"ContainerDied","Data":"9370c5745888934d7eaa524156e7b23861ec199211dc3d49521f028841ea3f51"} Dec 03 13:52:27 crc kubenswrapper[4677]: I1203 13:52:27.074421 4677 scope.go:117] "RemoveContainer" containerID="9370c5745888934d7eaa524156e7b23861ec199211dc3d49521f028841ea3f51" Dec 03 13:52:27 crc kubenswrapper[4677]: I1203 13:52:27.438102 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 13:52:28 crc kubenswrapper[4677]: I1203 13:52:28.274026 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:52:28 crc kubenswrapper[4677]: I1203 13:52:28.274361 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:52:29 crc kubenswrapper[4677]: I1203 13:52:29.104315 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" event={"ID":"9dc01e2f-388f-41a2-a139-5d251ad6cda8","Type":"ContainerStarted","Data":"d7a4e89157b13bb7b09d57dbfd96091406ea13a196ab64263f00b6ec6f2de879"} Dec 03 13:52:29 crc kubenswrapper[4677]: I1203 13:52:29.104570 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:52:29 crc kubenswrapper[4677]: I1203 13:52:29.108788 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:52:29 crc kubenswrapper[4677]: I1203 13:52:29.316742 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 13:52:30 crc kubenswrapper[4677]: I1203 13:52:30.070531 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 13:52:31 crc kubenswrapper[4677]: I1203 13:52:31.790453 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 13:52:32 crc kubenswrapper[4677]: I1203 13:52:32.053535 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4sxv"] Dec 03 13:52:32 crc kubenswrapper[4677]: I1203 13:52:32.053805 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w4sxv" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="registry-server" containerID="cri-o://ad0e62d1ac27c30b090514864fb494bdfcd7b343abb85ec53f276898d3597532" gracePeriod=2 Dec 03 13:52:32 crc kubenswrapper[4677]: I1203 13:52:32.063283 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 13:52:34 crc kubenswrapper[4677]: I1203 13:52:34.417211 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 13:52:35 crc kubenswrapper[4677]: I1203 13:52:35.142511 4677 generic.go:334] "Generic (PLEG): container finished" podID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerID="ad0e62d1ac27c30b090514864fb494bdfcd7b343abb85ec53f276898d3597532" exitCode=0 Dec 03 13:52:35 crc kubenswrapper[4677]: I1203 13:52:35.142640 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4sxv" event={"ID":"04d74a24-b7f5-4026-bea7-7eb2caef99ba","Type":"ContainerDied","Data":"ad0e62d1ac27c30b090514864fb494bdfcd7b343abb85ec53f276898d3597532"} Dec 03 13:52:35 crc kubenswrapper[4677]: I1203 13:52:35.970540 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 13:52:36 crc kubenswrapper[4677]: I1203 13:52:36.540889 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 13:52:36 crc kubenswrapper[4677]: I1203 13:52:36.993484 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.142679 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7wgt\" (UniqueName: \"kubernetes.io/projected/04d74a24-b7f5-4026-bea7-7eb2caef99ba-kube-api-access-p7wgt\") pod \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.143054 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-utilities\") pod \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.143083 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-catalog-content\") pod \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\" (UID: \"04d74a24-b7f5-4026-bea7-7eb2caef99ba\") " Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.144638 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-utilities" (OuterVolumeSpecName: "utilities") pod "04d74a24-b7f5-4026-bea7-7eb2caef99ba" (UID: "04d74a24-b7f5-4026-bea7-7eb2caef99ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.153000 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d74a24-b7f5-4026-bea7-7eb2caef99ba-kube-api-access-p7wgt" (OuterVolumeSpecName: "kube-api-access-p7wgt") pod "04d74a24-b7f5-4026-bea7-7eb2caef99ba" (UID: "04d74a24-b7f5-4026-bea7-7eb2caef99ba"). InnerVolumeSpecName "kube-api-access-p7wgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.161638 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4sxv" event={"ID":"04d74a24-b7f5-4026-bea7-7eb2caef99ba","Type":"ContainerDied","Data":"a2ead7f0896779c1fd73a8cabad6366b7326565034ecea1dfd9e397628fb351b"} Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.161753 4677 scope.go:117] "RemoveContainer" containerID="ad0e62d1ac27c30b090514864fb494bdfcd7b343abb85ec53f276898d3597532" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.162073 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4sxv" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.183484 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04d74a24-b7f5-4026-bea7-7eb2caef99ba" (UID: "04d74a24-b7f5-4026-bea7-7eb2caef99ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.184860 4677 scope.go:117] "RemoveContainer" containerID="4e701f20f53836c75b78740cabc3e29706f3ad17bb3ad377069c9b305edd4113" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.206720 4677 scope.go:117] "RemoveContainer" containerID="52fdfa64d3010fcae3d5a8ea0b4564c38dce6d3fd68fbc163b2fcccf59ee827a" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.244684 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.244742 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d74a24-b7f5-4026-bea7-7eb2caef99ba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.244760 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7wgt\" (UniqueName: \"kubernetes.io/projected/04d74a24-b7f5-4026-bea7-7eb2caef99ba-kube-api-access-p7wgt\") on node \"crc\" DevicePath \"\"" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.501983 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4sxv"] Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.506604 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4sxv"] Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.888871 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 13:52:37 crc kubenswrapper[4677]: I1203 13:52:37.987417 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" path="/var/lib/kubelet/pods/04d74a24-b7f5-4026-bea7-7eb2caef99ba/volumes" Dec 03 13:52:38 crc kubenswrapper[4677]: I1203 13:52:38.401980 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 13:52:40 crc kubenswrapper[4677]: I1203 13:52:40.576663 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 13:52:40 crc kubenswrapper[4677]: I1203 13:52:40.613132 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 13:52:42 crc kubenswrapper[4677]: I1203 13:52:42.066896 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 13:52:43 crc kubenswrapper[4677]: I1203 13:52:43.519780 4677 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 13:52:58 crc kubenswrapper[4677]: I1203 13:52:58.365942 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ppzxd"] Dec 03 13:52:58 crc kubenswrapper[4677]: I1203 13:52:58.366737 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" podUID="b22d27e8-6223-4370-9c02-ef73ec251631" containerName="controller-manager" containerID="cri-o://3fd18eb1a7faa05f3873f341a4ddd55df8fa892b2255f0516d12a3b24b1bfdce" gracePeriod=30 Dec 03 13:52:58 crc kubenswrapper[4677]: I1203 13:52:58.501158 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj"] Dec 03 13:52:58 crc kubenswrapper[4677]: I1203 13:52:58.501393 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" podUID="84259910-9b22-4957-ad88-7d00052ee57c" containerName="route-controller-manager" containerID="cri-o://95544b6169ec2ca29b515fe97accf1326d91c64bc3fef16856e54a394199b0d7" gracePeriod=30 Dec 03 13:52:59 crc kubenswrapper[4677]: I1203 13:52:59.313694 4677 generic.go:334] "Generic (PLEG): container finished" podID="84259910-9b22-4957-ad88-7d00052ee57c" containerID="95544b6169ec2ca29b515fe97accf1326d91c64bc3fef16856e54a394199b0d7" exitCode=0 Dec 03 13:52:59 crc kubenswrapper[4677]: I1203 13:52:59.313809 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" event={"ID":"84259910-9b22-4957-ad88-7d00052ee57c","Type":"ContainerDied","Data":"95544b6169ec2ca29b515fe97accf1326d91c64bc3fef16856e54a394199b0d7"} Dec 03 13:52:59 crc kubenswrapper[4677]: I1203 13:52:59.315312 4677 generic.go:334] "Generic (PLEG): container finished" podID="b22d27e8-6223-4370-9c02-ef73ec251631" containerID="3fd18eb1a7faa05f3873f341a4ddd55df8fa892b2255f0516d12a3b24b1bfdce" exitCode=0 Dec 03 13:52:59 crc kubenswrapper[4677]: I1203 13:52:59.315357 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" event={"ID":"b22d27e8-6223-4370-9c02-ef73ec251631","Type":"ContainerDied","Data":"3fd18eb1a7faa05f3873f341a4ddd55df8fa892b2255f0516d12a3b24b1bfdce"} Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.127674 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.161597 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-kszs9"] Dec 03 13:53:00 crc kubenswrapper[4677]: E1203 13:53:00.161859 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="extract-utilities" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.161881 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="extract-utilities" Dec 03 13:53:00 crc kubenswrapper[4677]: E1203 13:53:00.161895 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.161903 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 13:53:00 crc kubenswrapper[4677]: E1203 13:53:00.161917 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b22d27e8-6223-4370-9c02-ef73ec251631" containerName="controller-manager" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.161925 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b22d27e8-6223-4370-9c02-ef73ec251631" containerName="controller-manager" Dec 03 13:53:00 crc kubenswrapper[4677]: E1203 13:53:00.161938 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="extract-content" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.161977 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="extract-content" Dec 03 13:53:00 crc kubenswrapper[4677]: E1203 13:53:00.162002 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="registry-server" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.162011 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="registry-server" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.162174 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b22d27e8-6223-4370-9c02-ef73ec251631" containerName="controller-manager" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.162188 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d74a24-b7f5-4026-bea7-7eb2caef99ba" containerName="registry-server" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.162198 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.162632 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.178814 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-kszs9"] Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.184748 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.236541 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-client-ca\") pod \"b22d27e8-6223-4370-9c02-ef73ec251631\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.238392 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-proxy-ca-bundles\") pod \"b22d27e8-6223-4370-9c02-ef73ec251631\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.238582 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22d27e8-6223-4370-9c02-ef73ec251631-serving-cert\") pod \"b22d27e8-6223-4370-9c02-ef73ec251631\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.238621 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84259910-9b22-4957-ad88-7d00052ee57c-serving-cert\") pod \"84259910-9b22-4957-ad88-7d00052ee57c\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.238660 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-config\") pod \"b22d27e8-6223-4370-9c02-ef73ec251631\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.238702 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lztcv\" (UniqueName: \"kubernetes.io/projected/84259910-9b22-4957-ad88-7d00052ee57c-kube-api-access-lztcv\") pod \"84259910-9b22-4957-ad88-7d00052ee57c\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.238743 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n2rl\" (UniqueName: \"kubernetes.io/projected/b22d27e8-6223-4370-9c02-ef73ec251631-kube-api-access-5n2rl\") pod \"b22d27e8-6223-4370-9c02-ef73ec251631\" (UID: \"b22d27e8-6223-4370-9c02-ef73ec251631\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.238820 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-client-ca\") pod \"84259910-9b22-4957-ad88-7d00052ee57c\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.239021 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b22d27e8-6223-4370-9c02-ef73ec251631" (UID: "b22d27e8-6223-4370-9c02-ef73ec251631"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.239150 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-client-ca" (OuterVolumeSpecName: "client-ca") pod "b22d27e8-6223-4370-9c02-ef73ec251631" (UID: "b22d27e8-6223-4370-9c02-ef73ec251631"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.239872 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-config" (OuterVolumeSpecName: "config") pod "b22d27e8-6223-4370-9c02-ef73ec251631" (UID: "b22d27e8-6223-4370-9c02-ef73ec251631"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.239986 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-client-ca" (OuterVolumeSpecName: "client-ca") pod "84259910-9b22-4957-ad88-7d00052ee57c" (UID: "84259910-9b22-4957-ad88-7d00052ee57c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.240204 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-config\") pod \"84259910-9b22-4957-ad88-7d00052ee57c\" (UID: \"84259910-9b22-4957-ad88-7d00052ee57c\") " Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.240136 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-config" (OuterVolumeSpecName: "config") pod "84259910-9b22-4957-ad88-7d00052ee57c" (UID: "84259910-9b22-4957-ad88-7d00052ee57c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.240610 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12914d8a-4123-4b58-b250-3026bf2c903d-serving-cert\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.240664 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-config\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.240696 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-proxy-ca-bundles\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.240736 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-client-ca\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.241079 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knd2z\" (UniqueName: \"kubernetes.io/projected/12914d8a-4123-4b58-b250-3026bf2c903d-kube-api-access-knd2z\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.241248 4677 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.241783 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84259910-9b22-4957-ad88-7d00052ee57c-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.241959 4677 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.242053 4677 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.242143 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b22d27e8-6223-4370-9c02-ef73ec251631-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.245856 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b22d27e8-6223-4370-9c02-ef73ec251631-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b22d27e8-6223-4370-9c02-ef73ec251631" (UID: "b22d27e8-6223-4370-9c02-ef73ec251631"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.245997 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b22d27e8-6223-4370-9c02-ef73ec251631-kube-api-access-5n2rl" (OuterVolumeSpecName: "kube-api-access-5n2rl") pod "b22d27e8-6223-4370-9c02-ef73ec251631" (UID: "b22d27e8-6223-4370-9c02-ef73ec251631"). InnerVolumeSpecName "kube-api-access-5n2rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.248031 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84259910-9b22-4957-ad88-7d00052ee57c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "84259910-9b22-4957-ad88-7d00052ee57c" (UID: "84259910-9b22-4957-ad88-7d00052ee57c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.248694 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84259910-9b22-4957-ad88-7d00052ee57c-kube-api-access-lztcv" (OuterVolumeSpecName: "kube-api-access-lztcv") pod "84259910-9b22-4957-ad88-7d00052ee57c" (UID: "84259910-9b22-4957-ad88-7d00052ee57c"). InnerVolumeSpecName "kube-api-access-lztcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.325366 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" event={"ID":"b22d27e8-6223-4370-9c02-ef73ec251631","Type":"ContainerDied","Data":"efad9c818ccff235f6b00795f3679034b3617c6741a5b5b6be28a8c69d2f2445"} Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.325415 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ppzxd" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.325431 4677 scope.go:117] "RemoveContainer" containerID="3fd18eb1a7faa05f3873f341a4ddd55df8fa892b2255f0516d12a3b24b1bfdce" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.327499 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" event={"ID":"84259910-9b22-4957-ad88-7d00052ee57c","Type":"ContainerDied","Data":"0aa3ba1806b4b598a7d990bf09893d1183aec1454270660f3d39698065315925"} Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.327572 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.340727 4677 scope.go:117] "RemoveContainer" containerID="95544b6169ec2ca29b515fe97accf1326d91c64bc3fef16856e54a394199b0d7" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.342800 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-client-ca\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.342901 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knd2z\" (UniqueName: \"kubernetes.io/projected/12914d8a-4123-4b58-b250-3026bf2c903d-kube-api-access-knd2z\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.343152 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12914d8a-4123-4b58-b250-3026bf2c903d-serving-cert\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.343912 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-config\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.343981 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-proxy-ca-bundles\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.344139 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b22d27e8-6223-4370-9c02-ef73ec251631-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.344160 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84259910-9b22-4957-ad88-7d00052ee57c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.344172 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lztcv\" (UniqueName: \"kubernetes.io/projected/84259910-9b22-4957-ad88-7d00052ee57c-kube-api-access-lztcv\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.344183 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n2rl\" (UniqueName: \"kubernetes.io/projected/b22d27e8-6223-4370-9c02-ef73ec251631-kube-api-access-5n2rl\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.345192 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-config\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.345619 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-proxy-ca-bundles\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.345744 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-client-ca\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.347430 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12914d8a-4123-4b58-b250-3026bf2c903d-serving-cert\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.354052 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ppzxd"] Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.358717 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ppzxd"] Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.363626 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knd2z\" (UniqueName: \"kubernetes.io/projected/12914d8a-4123-4b58-b250-3026bf2c903d-kube-api-access-knd2z\") pod \"controller-manager-55b446ff67-kszs9\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.365612 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj"] Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.370088 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwsjj"] Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.503211 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:00 crc kubenswrapper[4677]: I1203 13:53:00.674688 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-kszs9"] Dec 03 13:53:01 crc kubenswrapper[4677]: I1203 13:53:01.336464 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" event={"ID":"12914d8a-4123-4b58-b250-3026bf2c903d","Type":"ContainerStarted","Data":"b09c766fdd45b57134259458d499e3323f03aee09df8c9c24cc90c370ee57f01"} Dec 03 13:53:01 crc kubenswrapper[4677]: I1203 13:53:01.336797 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" event={"ID":"12914d8a-4123-4b58-b250-3026bf2c903d","Type":"ContainerStarted","Data":"e9137cd88842a30438636597ef097b1fe52bfcb823a2b870278b3a4eb7c5c6d4"} Dec 03 13:53:01 crc kubenswrapper[4677]: I1203 13:53:01.336815 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:01 crc kubenswrapper[4677]: I1203 13:53:01.341774 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:01 crc kubenswrapper[4677]: I1203 13:53:01.353240 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" podStartSLOduration=3.353220878 podStartE2EDuration="3.353220878s" podCreationTimestamp="2025-12-03 13:52:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:01.353042493 +0000 UTC m=+372.099374958" watchObservedRunningTime="2025-12-03 13:53:01.353220878 +0000 UTC m=+372.099553333" Dec 03 13:53:01 crc kubenswrapper[4677]: I1203 13:53:01.982589 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84259910-9b22-4957-ad88-7d00052ee57c" path="/var/lib/kubelet/pods/84259910-9b22-4957-ad88-7d00052ee57c/volumes" Dec 03 13:53:01 crc kubenswrapper[4677]: I1203 13:53:01.983100 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b22d27e8-6223-4370-9c02-ef73ec251631" path="/var/lib/kubelet/pods/b22d27e8-6223-4370-9c02-ef73ec251631/volumes" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.293915 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7"] Dec 03 13:53:02 crc kubenswrapper[4677]: E1203 13:53:02.294161 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84259910-9b22-4957-ad88-7d00052ee57c" containerName="route-controller-manager" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.294174 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="84259910-9b22-4957-ad88-7d00052ee57c" containerName="route-controller-manager" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.294270 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="84259910-9b22-4957-ad88-7d00052ee57c" containerName="route-controller-manager" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.294594 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.297358 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.298084 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.298209 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.299319 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.299485 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.299629 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.307184 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7"] Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.368848 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-client-ca\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.368908 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt82j\" (UniqueName: \"kubernetes.io/projected/c6feadb1-739a-470e-b876-9b934d924490-kube-api-access-zt82j\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.368968 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6feadb1-739a-470e-b876-9b934d924490-serving-cert\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.369035 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-config\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.469749 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6feadb1-739a-470e-b876-9b934d924490-serving-cert\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.469820 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-config\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.469851 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-client-ca\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.469887 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt82j\" (UniqueName: \"kubernetes.io/projected/c6feadb1-739a-470e-b876-9b934d924490-kube-api-access-zt82j\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.470855 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-client-ca\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.471471 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-config\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.486926 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt82j\" (UniqueName: \"kubernetes.io/projected/c6feadb1-739a-470e-b876-9b934d924490-kube-api-access-zt82j\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.487083 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6feadb1-739a-470e-b876-9b934d924490-serving-cert\") pod \"route-controller-manager-646bddd5f-qftk7\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:02 crc kubenswrapper[4677]: I1203 13:53:02.616695 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:03 crc kubenswrapper[4677]: I1203 13:53:03.032428 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7"] Dec 03 13:53:03 crc kubenswrapper[4677]: W1203 13:53:03.040157 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6feadb1_739a_470e_b876_9b934d924490.slice/crio-236161d54505eb616ae6dca8a69c7711435e24e93bc5bde7ec0dbcd6f77375fa WatchSource:0}: Error finding container 236161d54505eb616ae6dca8a69c7711435e24e93bc5bde7ec0dbcd6f77375fa: Status 404 returned error can't find the container with id 236161d54505eb616ae6dca8a69c7711435e24e93bc5bde7ec0dbcd6f77375fa Dec 03 13:53:03 crc kubenswrapper[4677]: I1203 13:53:03.362261 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" event={"ID":"c6feadb1-739a-470e-b876-9b934d924490","Type":"ContainerStarted","Data":"236161d54505eb616ae6dca8a69c7711435e24e93bc5bde7ec0dbcd6f77375fa"} Dec 03 13:53:04 crc kubenswrapper[4677]: I1203 13:53:04.369794 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" event={"ID":"c6feadb1-739a-470e-b876-9b934d924490","Type":"ContainerStarted","Data":"91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683"} Dec 03 13:53:04 crc kubenswrapper[4677]: I1203 13:53:04.372431 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:04 crc kubenswrapper[4677]: I1203 13:53:04.406060 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" podStartSLOduration=6.406029647 podStartE2EDuration="6.406029647s" podCreationTimestamp="2025-12-03 13:52:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:04.401306422 +0000 UTC m=+375.147638887" watchObservedRunningTime="2025-12-03 13:53:04.406029647 +0000 UTC m=+375.152362142" Dec 03 13:53:04 crc kubenswrapper[4677]: I1203 13:53:04.423053 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:08 crc kubenswrapper[4677]: I1203 13:53:08.437617 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:53:08 crc kubenswrapper[4677]: I1203 13:53:08.437951 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:53:09 crc kubenswrapper[4677]: I1203 13:53:09.490786 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbnmw"] Dec 03 13:53:09 crc kubenswrapper[4677]: I1203 13:53:09.491022 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dbnmw" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="registry-server" containerID="cri-o://1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0" gracePeriod=2 Dec 03 13:53:09 crc kubenswrapper[4677]: I1203 13:53:09.898280 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.073464 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-utilities\") pod \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.073520 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-catalog-content\") pod \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.074097 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mcns\" (UniqueName: \"kubernetes.io/projected/96689c77-5d8c-46c9-a87b-fc98bf5797ed-kube-api-access-5mcns\") pod \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\" (UID: \"96689c77-5d8c-46c9-a87b-fc98bf5797ed\") " Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.075656 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-utilities" (OuterVolumeSpecName: "utilities") pod "96689c77-5d8c-46c9-a87b-fc98bf5797ed" (UID: "96689c77-5d8c-46c9-a87b-fc98bf5797ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.085451 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96689c77-5d8c-46c9-a87b-fc98bf5797ed-kube-api-access-5mcns" (OuterVolumeSpecName: "kube-api-access-5mcns") pod "96689c77-5d8c-46c9-a87b-fc98bf5797ed" (UID: "96689c77-5d8c-46c9-a87b-fc98bf5797ed"). InnerVolumeSpecName "kube-api-access-5mcns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.118786 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96689c77-5d8c-46c9-a87b-fc98bf5797ed" (UID: "96689c77-5d8c-46c9-a87b-fc98bf5797ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.175894 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.175936 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96689c77-5d8c-46c9-a87b-fc98bf5797ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.175950 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mcns\" (UniqueName: \"kubernetes.io/projected/96689c77-5d8c-46c9-a87b-fc98bf5797ed-kube-api-access-5mcns\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.427994 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbnmw" event={"ID":"96689c77-5d8c-46c9-a87b-fc98bf5797ed","Type":"ContainerDied","Data":"1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0"} Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.428410 4677 scope.go:117] "RemoveContainer" containerID="1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.428003 4677 generic.go:334] "Generic (PLEG): container finished" podID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerID="1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0" exitCode=0 Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.428498 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbnmw" event={"ID":"96689c77-5d8c-46c9-a87b-fc98bf5797ed","Type":"ContainerDied","Data":"ee3d4cd03206327d2374fb88f1c248be17229b8981bb7286094232d5c5d8e928"} Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.428034 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbnmw" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.449101 4677 scope.go:117] "RemoveContainer" containerID="c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.456642 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbnmw"] Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.474200 4677 scope.go:117] "RemoveContainer" containerID="f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.479676 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dbnmw"] Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.493622 4677 scope.go:117] "RemoveContainer" containerID="1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0" Dec 03 13:53:10 crc kubenswrapper[4677]: E1203 13:53:10.494117 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0\": container with ID starting with 1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0 not found: ID does not exist" containerID="1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.494146 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0"} err="failed to get container status \"1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0\": rpc error: code = NotFound desc = could not find container \"1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0\": container with ID starting with 1321d70ba771a4ccf4b5cbac5c30eabaade2745ebe768affd4ae85151cba07f0 not found: ID does not exist" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.494167 4677 scope.go:117] "RemoveContainer" containerID="c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b" Dec 03 13:53:10 crc kubenswrapper[4677]: E1203 13:53:10.494874 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b\": container with ID starting with c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b not found: ID does not exist" containerID="c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.494893 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b"} err="failed to get container status \"c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b\": rpc error: code = NotFound desc = could not find container \"c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b\": container with ID starting with c6df77ddaa9a54322e975129ecab5c4d0aab9edba64cc9e12d188c75d921399b not found: ID does not exist" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.494907 4677 scope.go:117] "RemoveContainer" containerID="f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb" Dec 03 13:53:10 crc kubenswrapper[4677]: E1203 13:53:10.495256 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb\": container with ID starting with f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb not found: ID does not exist" containerID="f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb" Dec 03 13:53:10 crc kubenswrapper[4677]: I1203 13:53:10.495273 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb"} err="failed to get container status \"f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb\": rpc error: code = NotFound desc = could not find container \"f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb\": container with ID starting with f1f87d9d460c8d0177a99c5d8519d7fb198a228d9d6115fa88b7235e8daa3bbb not found: ID does not exist" Dec 03 13:53:11 crc kubenswrapper[4677]: I1203 13:53:11.891246 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2qhnw"] Dec 03 13:53:11 crc kubenswrapper[4677]: I1203 13:53:11.891503 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2qhnw" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="registry-server" containerID="cri-o://b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db" gracePeriod=2 Dec 03 13:53:11 crc kubenswrapper[4677]: I1203 13:53:11.982898 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" path="/var/lib/kubelet/pods/96689c77-5d8c-46c9-a87b-fc98bf5797ed/volumes" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.268652 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.400725 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-utilities\") pod \"d9a210ab-e032-4834-bb1c-b9835456d960\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.400855 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-catalog-content\") pod \"d9a210ab-e032-4834-bb1c-b9835456d960\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.400905 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn4th\" (UniqueName: \"kubernetes.io/projected/d9a210ab-e032-4834-bb1c-b9835456d960-kube-api-access-wn4th\") pod \"d9a210ab-e032-4834-bb1c-b9835456d960\" (UID: \"d9a210ab-e032-4834-bb1c-b9835456d960\") " Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.401603 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-utilities" (OuterVolumeSpecName: "utilities") pod "d9a210ab-e032-4834-bb1c-b9835456d960" (UID: "d9a210ab-e032-4834-bb1c-b9835456d960"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.405658 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9a210ab-e032-4834-bb1c-b9835456d960-kube-api-access-wn4th" (OuterVolumeSpecName: "kube-api-access-wn4th") pod "d9a210ab-e032-4834-bb1c-b9835456d960" (UID: "d9a210ab-e032-4834-bb1c-b9835456d960"). InnerVolumeSpecName "kube-api-access-wn4th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.442451 4677 generic.go:334] "Generic (PLEG): container finished" podID="d9a210ab-e032-4834-bb1c-b9835456d960" containerID="b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db" exitCode=0 Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.442498 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhnw" event={"ID":"d9a210ab-e032-4834-bb1c-b9835456d960","Type":"ContainerDied","Data":"b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db"} Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.442526 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhnw" event={"ID":"d9a210ab-e032-4834-bb1c-b9835456d960","Type":"ContainerDied","Data":"6b8d92e5ed8e3e06524576d7098915ef9faf7d171530803d672edd3e79c43ab8"} Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.442503 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhnw" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.442547 4677 scope.go:117] "RemoveContainer" containerID="b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.461224 4677 scope.go:117] "RemoveContainer" containerID="1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.479748 4677 scope.go:117] "RemoveContainer" containerID="e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.497399 4677 scope.go:117] "RemoveContainer" containerID="b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db" Dec 03 13:53:12 crc kubenswrapper[4677]: E1203 13:53:12.498218 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db\": container with ID starting with b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db not found: ID does not exist" containerID="b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.498261 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db"} err="failed to get container status \"b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db\": rpc error: code = NotFound desc = could not find container \"b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db\": container with ID starting with b39f52cb6c036dcbaccbc10c38c9798ab0cea14a94e77f7f298d5ec837f8d2db not found: ID does not exist" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.498291 4677 scope.go:117] "RemoveContainer" containerID="1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a" Dec 03 13:53:12 crc kubenswrapper[4677]: E1203 13:53:12.498738 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a\": container with ID starting with 1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a not found: ID does not exist" containerID="1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.498772 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a"} err="failed to get container status \"1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a\": rpc error: code = NotFound desc = could not find container \"1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a\": container with ID starting with 1d66f10236f257513bf06eabbaa70d9d895466401bb9f630ffc30f7737c7457a not found: ID does not exist" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.498790 4677 scope.go:117] "RemoveContainer" containerID="e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb" Dec 03 13:53:12 crc kubenswrapper[4677]: E1203 13:53:12.499351 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb\": container with ID starting with e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb not found: ID does not exist" containerID="e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.499418 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb"} err="failed to get container status \"e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb\": rpc error: code = NotFound desc = could not find container \"e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb\": container with ID starting with e5f480f6c7ccabc038729fbaef7ee9c42715fd0586343aeed58dbff893742aeb not found: ID does not exist" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.502398 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn4th\" (UniqueName: \"kubernetes.io/projected/d9a210ab-e032-4834-bb1c-b9835456d960-kube-api-access-wn4th\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.502422 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.516830 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9a210ab-e032-4834-bb1c-b9835456d960" (UID: "d9a210ab-e032-4834-bb1c-b9835456d960"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.605169 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a210ab-e032-4834-bb1c-b9835456d960-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.767689 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2qhnw"] Dec 03 13:53:12 crc kubenswrapper[4677]: I1203 13:53:12.773453 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2qhnw"] Dec 03 13:53:13 crc kubenswrapper[4677]: I1203 13:53:13.985908 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" path="/var/lib/kubelet/pods/d9a210ab-e032-4834-bb1c-b9835456d960/volumes" Dec 03 13:53:18 crc kubenswrapper[4677]: I1203 13:53:18.379190 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7"] Dec 03 13:53:18 crc kubenswrapper[4677]: I1203 13:53:18.379422 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" podUID="c6feadb1-739a-470e-b876-9b934d924490" containerName="route-controller-manager" containerID="cri-o://91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683" gracePeriod=30 Dec 03 13:53:18 crc kubenswrapper[4677]: I1203 13:53:18.842139 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:18 crc kubenswrapper[4677]: I1203 13:53:18.996649 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-config" (OuterVolumeSpecName: "config") pod "c6feadb1-739a-470e-b876-9b934d924490" (UID: "c6feadb1-739a-470e-b876-9b934d924490"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:18 crc kubenswrapper[4677]: I1203 13:53:18.996744 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-config\") pod \"c6feadb1-739a-470e-b876-9b934d924490\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:18.998128 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6feadb1-739a-470e-b876-9b934d924490-serving-cert\") pod \"c6feadb1-739a-470e-b876-9b934d924490\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:18.998268 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt82j\" (UniqueName: \"kubernetes.io/projected/c6feadb1-739a-470e-b876-9b934d924490-kube-api-access-zt82j\") pod \"c6feadb1-739a-470e-b876-9b934d924490\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:18.998310 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-client-ca\") pod \"c6feadb1-739a-470e-b876-9b934d924490\" (UID: \"c6feadb1-739a-470e-b876-9b934d924490\") " Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:18.998666 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:18.999238 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-client-ca" (OuterVolumeSpecName: "client-ca") pod "c6feadb1-739a-470e-b876-9b934d924490" (UID: "c6feadb1-739a-470e-b876-9b934d924490"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.004206 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6feadb1-739a-470e-b876-9b934d924490-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c6feadb1-739a-470e-b876-9b934d924490" (UID: "c6feadb1-739a-470e-b876-9b934d924490"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.009181 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6feadb1-739a-470e-b876-9b934d924490-kube-api-access-zt82j" (OuterVolumeSpecName: "kube-api-access-zt82j") pod "c6feadb1-739a-470e-b876-9b934d924490" (UID: "c6feadb1-739a-470e-b876-9b934d924490"). InnerVolumeSpecName "kube-api-access-zt82j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.099485 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt82j\" (UniqueName: \"kubernetes.io/projected/c6feadb1-739a-470e-b876-9b934d924490-kube-api-access-zt82j\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.099520 4677 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6feadb1-739a-470e-b876-9b934d924490-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.099528 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6feadb1-739a-470e-b876-9b934d924490-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.481092 4677 generic.go:334] "Generic (PLEG): container finished" podID="c6feadb1-739a-470e-b876-9b934d924490" containerID="91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683" exitCode=0 Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.481140 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" event={"ID":"c6feadb1-739a-470e-b876-9b934d924490","Type":"ContainerDied","Data":"91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683"} Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.481169 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" event={"ID":"c6feadb1-739a-470e-b876-9b934d924490","Type":"ContainerDied","Data":"236161d54505eb616ae6dca8a69c7711435e24e93bc5bde7ec0dbcd6f77375fa"} Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.481188 4677 scope.go:117] "RemoveContainer" containerID="91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.481302 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.506897 4677 scope.go:117] "RemoveContainer" containerID="91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683" Dec 03 13:53:19 crc kubenswrapper[4677]: E1203 13:53:19.507746 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683\": container with ID starting with 91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683 not found: ID does not exist" containerID="91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.507843 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683"} err="failed to get container status \"91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683\": rpc error: code = NotFound desc = could not find container \"91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683\": container with ID starting with 91cf4369760edefb12d8f05f62c7ad012a61357ce3bdec4a32bea49314159683 not found: ID does not exist" Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.525724 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7"] Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.529590 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-qftk7"] Dec 03 13:53:19 crc kubenswrapper[4677]: I1203 13:53:19.981856 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6feadb1-739a-470e-b876-9b934d924490" path="/var/lib/kubelet/pods/c6feadb1-739a-470e-b876-9b934d924490/volumes" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309136 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98"] Dec 03 13:53:20 crc kubenswrapper[4677]: E1203 13:53:20.309433 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="extract-utilities" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309447 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="extract-utilities" Dec 03 13:53:20 crc kubenswrapper[4677]: E1203 13:53:20.309459 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="registry-server" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309466 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="registry-server" Dec 03 13:53:20 crc kubenswrapper[4677]: E1203 13:53:20.309481 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="extract-content" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309489 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="extract-content" Dec 03 13:53:20 crc kubenswrapper[4677]: E1203 13:53:20.309501 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6feadb1-739a-470e-b876-9b934d924490" containerName="route-controller-manager" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309508 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6feadb1-739a-470e-b876-9b934d924490" containerName="route-controller-manager" Dec 03 13:53:20 crc kubenswrapper[4677]: E1203 13:53:20.309517 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="extract-content" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309523 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="extract-content" Dec 03 13:53:20 crc kubenswrapper[4677]: E1203 13:53:20.309534 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="extract-utilities" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309540 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="extract-utilities" Dec 03 13:53:20 crc kubenswrapper[4677]: E1203 13:53:20.309549 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="registry-server" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309555 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="registry-server" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309659 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6feadb1-739a-470e-b876-9b934d924490" containerName="route-controller-manager" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309673 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="96689c77-5d8c-46c9-a87b-fc98bf5797ed" containerName="registry-server" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.309689 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9a210ab-e032-4834-bb1c-b9835456d960" containerName="registry-server" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.310152 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.312119 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.313047 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.313145 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.313268 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.313366 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.313538 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.314971 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8b1be0-e4f3-48a9-a8af-365480f92351-serving-cert\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.315020 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc8b1be0-e4f3-48a9-a8af-365480f92351-client-ca\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.315074 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8b1be0-e4f3-48a9-a8af-365480f92351-config\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.315130 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2gs9\" (UniqueName: \"kubernetes.io/projected/fc8b1be0-e4f3-48a9-a8af-365480f92351-kube-api-access-k2gs9\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.315554 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98"] Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.416400 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8b1be0-e4f3-48a9-a8af-365480f92351-serving-cert\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.416507 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc8b1be0-e4f3-48a9-a8af-365480f92351-client-ca\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.416623 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8b1be0-e4f3-48a9-a8af-365480f92351-config\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.416707 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2gs9\" (UniqueName: \"kubernetes.io/projected/fc8b1be0-e4f3-48a9-a8af-365480f92351-kube-api-access-k2gs9\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.418044 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc8b1be0-e4f3-48a9-a8af-365480f92351-client-ca\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.418323 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8b1be0-e4f3-48a9-a8af-365480f92351-config\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.422038 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8b1be0-e4f3-48a9-a8af-365480f92351-serving-cert\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.442549 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2gs9\" (UniqueName: \"kubernetes.io/projected/fc8b1be0-e4f3-48a9-a8af-365480f92351-kube-api-access-k2gs9\") pod \"route-controller-manager-6748b696b6-jds98\" (UID: \"fc8b1be0-e4f3-48a9-a8af-365480f92351\") " pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:20 crc kubenswrapper[4677]: I1203 13:53:20.627006 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:21 crc kubenswrapper[4677]: I1203 13:53:21.074626 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98"] Dec 03 13:53:21 crc kubenswrapper[4677]: W1203 13:53:21.088131 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc8b1be0_e4f3_48a9_a8af_365480f92351.slice/crio-72bd5c9a701c55b57a27d1eb65425b7e787854b3cd50a21b30454461e090d382 WatchSource:0}: Error finding container 72bd5c9a701c55b57a27d1eb65425b7e787854b3cd50a21b30454461e090d382: Status 404 returned error can't find the container with id 72bd5c9a701c55b57a27d1eb65425b7e787854b3cd50a21b30454461e090d382 Dec 03 13:53:21 crc kubenswrapper[4677]: I1203 13:53:21.493978 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" event={"ID":"fc8b1be0-e4f3-48a9-a8af-365480f92351","Type":"ContainerStarted","Data":"d758c866e506ab7847f3f70c8ecb7b0eed697eb027a7dd1093601bf8283543fa"} Dec 03 13:53:21 crc kubenswrapper[4677]: I1203 13:53:21.494220 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" event={"ID":"fc8b1be0-e4f3-48a9-a8af-365480f92351","Type":"ContainerStarted","Data":"72bd5c9a701c55b57a27d1eb65425b7e787854b3cd50a21b30454461e090d382"} Dec 03 13:53:21 crc kubenswrapper[4677]: I1203 13:53:21.494544 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:21 crc kubenswrapper[4677]: I1203 13:53:21.529991 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" podStartSLOduration=3.529974412 podStartE2EDuration="3.529974412s" podCreationTimestamp="2025-12-03 13:53:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:21.529658773 +0000 UTC m=+392.275991248" watchObservedRunningTime="2025-12-03 13:53:21.529974412 +0000 UTC m=+392.276306867" Dec 03 13:53:21 crc kubenswrapper[4677]: I1203 13:53:21.736525 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6748b696b6-jds98" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.363359 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x9htr"] Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.364679 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.388188 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x9htr"] Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524396 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-registry-tls\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524482 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6gwh\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-kube-api-access-z6gwh\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524504 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/066f25c8-d604-405a-ac2e-cbcaa0d76772-registry-certificates\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524539 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/066f25c8-d604-405a-ac2e-cbcaa0d76772-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524560 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/066f25c8-d604-405a-ac2e-cbcaa0d76772-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524609 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524637 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-bound-sa-token\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.524704 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/066f25c8-d604-405a-ac2e-cbcaa0d76772-trusted-ca\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.547652 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.625592 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/066f25c8-d604-405a-ac2e-cbcaa0d76772-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.625643 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/066f25c8-d604-405a-ac2e-cbcaa0d76772-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.625683 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-bound-sa-token\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.625736 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/066f25c8-d604-405a-ac2e-cbcaa0d76772-trusted-ca\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.625792 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-registry-tls\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.625816 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6gwh\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-kube-api-access-z6gwh\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.625844 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/066f25c8-d604-405a-ac2e-cbcaa0d76772-registry-certificates\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.626411 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/066f25c8-d604-405a-ac2e-cbcaa0d76772-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.627213 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/066f25c8-d604-405a-ac2e-cbcaa0d76772-registry-certificates\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.627848 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/066f25c8-d604-405a-ac2e-cbcaa0d76772-trusted-ca\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.639630 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/066f25c8-d604-405a-ac2e-cbcaa0d76772-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.639749 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-registry-tls\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.645099 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-bound-sa-token\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.645405 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6gwh\" (UniqueName: \"kubernetes.io/projected/066f25c8-d604-405a-ac2e-cbcaa0d76772-kube-api-access-z6gwh\") pod \"image-registry-66df7c8f76-x9htr\" (UID: \"066f25c8-d604-405a-ac2e-cbcaa0d76772\") " pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:35 crc kubenswrapper[4677]: I1203 13:53:35.686468 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:36 crc kubenswrapper[4677]: I1203 13:53:36.177669 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x9htr"] Dec 03 13:53:36 crc kubenswrapper[4677]: W1203 13:53:36.179185 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod066f25c8_d604_405a_ac2e_cbcaa0d76772.slice/crio-59fb88939c5f70122495bcdb383b872bbaffd1779afa5a171d05b0ecb1457d6d WatchSource:0}: Error finding container 59fb88939c5f70122495bcdb383b872bbaffd1779afa5a171d05b0ecb1457d6d: Status 404 returned error can't find the container with id 59fb88939c5f70122495bcdb383b872bbaffd1779afa5a171d05b0ecb1457d6d Dec 03 13:53:36 crc kubenswrapper[4677]: I1203 13:53:36.584541 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" event={"ID":"066f25c8-d604-405a-ac2e-cbcaa0d76772","Type":"ContainerStarted","Data":"2e3c34352a6b3235a0d86cddfc1ec79abe81948851b8509013b9af1944af3e7e"} Dec 03 13:53:36 crc kubenswrapper[4677]: I1203 13:53:36.584878 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" event={"ID":"066f25c8-d604-405a-ac2e-cbcaa0d76772","Type":"ContainerStarted","Data":"59fb88939c5f70122495bcdb383b872bbaffd1779afa5a171d05b0ecb1457d6d"} Dec 03 13:53:36 crc kubenswrapper[4677]: I1203 13:53:36.584898 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.370876 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" podStartSLOduration=3.370854927 podStartE2EDuration="3.370854927s" podCreationTimestamp="2025-12-03 13:53:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:36.610911628 +0000 UTC m=+407.357244103" watchObservedRunningTime="2025-12-03 13:53:38.370854927 +0000 UTC m=+409.117187382" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.373350 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-kszs9"] Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.373903 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" podUID="12914d8a-4123-4b58-b250-3026bf2c903d" containerName="controller-manager" containerID="cri-o://b09c766fdd45b57134259458d499e3323f03aee09df8c9c24cc90c370ee57f01" gracePeriod=30 Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.437672 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.437759 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.598159 4677 generic.go:334] "Generic (PLEG): container finished" podID="12914d8a-4123-4b58-b250-3026bf2c903d" containerID="b09c766fdd45b57134259458d499e3323f03aee09df8c9c24cc90c370ee57f01" exitCode=0 Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.598240 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" event={"ID":"12914d8a-4123-4b58-b250-3026bf2c903d","Type":"ContainerDied","Data":"b09c766fdd45b57134259458d499e3323f03aee09df8c9c24cc90c370ee57f01"} Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.744602 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.867430 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-client-ca\") pod \"12914d8a-4123-4b58-b250-3026bf2c903d\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.867488 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knd2z\" (UniqueName: \"kubernetes.io/projected/12914d8a-4123-4b58-b250-3026bf2c903d-kube-api-access-knd2z\") pod \"12914d8a-4123-4b58-b250-3026bf2c903d\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.867517 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-config\") pod \"12914d8a-4123-4b58-b250-3026bf2c903d\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.867539 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12914d8a-4123-4b58-b250-3026bf2c903d-serving-cert\") pod \"12914d8a-4123-4b58-b250-3026bf2c903d\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.867611 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-proxy-ca-bundles\") pod \"12914d8a-4123-4b58-b250-3026bf2c903d\" (UID: \"12914d8a-4123-4b58-b250-3026bf2c903d\") " Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.868626 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "12914d8a-4123-4b58-b250-3026bf2c903d" (UID: "12914d8a-4123-4b58-b250-3026bf2c903d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.868643 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-client-ca" (OuterVolumeSpecName: "client-ca") pod "12914d8a-4123-4b58-b250-3026bf2c903d" (UID: "12914d8a-4123-4b58-b250-3026bf2c903d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.869068 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-config" (OuterVolumeSpecName: "config") pod "12914d8a-4123-4b58-b250-3026bf2c903d" (UID: "12914d8a-4123-4b58-b250-3026bf2c903d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.873452 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12914d8a-4123-4b58-b250-3026bf2c903d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "12914d8a-4123-4b58-b250-3026bf2c903d" (UID: "12914d8a-4123-4b58-b250-3026bf2c903d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.873618 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12914d8a-4123-4b58-b250-3026bf2c903d-kube-api-access-knd2z" (OuterVolumeSpecName: "kube-api-access-knd2z") pod "12914d8a-4123-4b58-b250-3026bf2c903d" (UID: "12914d8a-4123-4b58-b250-3026bf2c903d"). InnerVolumeSpecName "kube-api-access-knd2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.968796 4677 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.968854 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knd2z\" (UniqueName: \"kubernetes.io/projected/12914d8a-4123-4b58-b250-3026bf2c903d-kube-api-access-knd2z\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.968874 4677 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12914d8a-4123-4b58-b250-3026bf2c903d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.968891 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:38 crc kubenswrapper[4677]: I1203 13:53:38.968911 4677 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12914d8a-4123-4b58-b250-3026bf2c903d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:39 crc kubenswrapper[4677]: I1203 13:53:39.605941 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" event={"ID":"12914d8a-4123-4b58-b250-3026bf2c903d","Type":"ContainerDied","Data":"e9137cd88842a30438636597ef097b1fe52bfcb823a2b870278b3a4eb7c5c6d4"} Dec 03 13:53:39 crc kubenswrapper[4677]: I1203 13:53:39.606004 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-kszs9" Dec 03 13:53:39 crc kubenswrapper[4677]: I1203 13:53:39.606025 4677 scope.go:117] "RemoveContainer" containerID="b09c766fdd45b57134259458d499e3323f03aee09df8c9c24cc90c370ee57f01" Dec 03 13:53:39 crc kubenswrapper[4677]: I1203 13:53:39.636902 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-kszs9"] Dec 03 13:53:39 crc kubenswrapper[4677]: I1203 13:53:39.641639 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-kszs9"] Dec 03 13:53:39 crc kubenswrapper[4677]: I1203 13:53:39.984200 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12914d8a-4123-4b58-b250-3026bf2c903d" path="/var/lib/kubelet/pods/12914d8a-4123-4b58-b250-3026bf2c903d/volumes" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.317232 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6589dc88cb-f294z"] Dec 03 13:53:40 crc kubenswrapper[4677]: E1203 13:53:40.317484 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12914d8a-4123-4b58-b250-3026bf2c903d" containerName="controller-manager" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.317502 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="12914d8a-4123-4b58-b250-3026bf2c903d" containerName="controller-manager" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.317635 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="12914d8a-4123-4b58-b250-3026bf2c903d" containerName="controller-manager" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.318119 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.322883 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.323009 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.323578 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.323616 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.323856 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.325414 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.341481 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.347012 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6589dc88cb-f294z"] Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.386160 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-proxy-ca-bundles\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.386455 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-config\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.386678 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-client-ca\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.386793 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c803dd49-87e9-424d-9728-c46bd96b36f9-serving-cert\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.386825 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chshn\" (UniqueName: \"kubernetes.io/projected/c803dd49-87e9-424d-9728-c46bd96b36f9-kube-api-access-chshn\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.487658 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c803dd49-87e9-424d-9728-c46bd96b36f9-serving-cert\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.487711 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chshn\" (UniqueName: \"kubernetes.io/projected/c803dd49-87e9-424d-9728-c46bd96b36f9-kube-api-access-chshn\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.487753 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-proxy-ca-bundles\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.487790 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-config\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.487823 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-client-ca\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.488757 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-client-ca\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.489261 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-config\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.489332 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c803dd49-87e9-424d-9728-c46bd96b36f9-proxy-ca-bundles\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.492731 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c803dd49-87e9-424d-9728-c46bd96b36f9-serving-cert\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.506917 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chshn\" (UniqueName: \"kubernetes.io/projected/c803dd49-87e9-424d-9728-c46bd96b36f9-kube-api-access-chshn\") pod \"controller-manager-6589dc88cb-f294z\" (UID: \"c803dd49-87e9-424d-9728-c46bd96b36f9\") " pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.642475 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:40 crc kubenswrapper[4677]: I1203 13:53:40.849557 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6589dc88cb-f294z"] Dec 03 13:53:41 crc kubenswrapper[4677]: I1203 13:53:41.619899 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" event={"ID":"c803dd49-87e9-424d-9728-c46bd96b36f9","Type":"ContainerStarted","Data":"40f198913afac49f5147c406d930345f50d492d9e6ab60ecce6de96fe42af8fd"} Dec 03 13:53:41 crc kubenswrapper[4677]: I1203 13:53:41.620300 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" event={"ID":"c803dd49-87e9-424d-9728-c46bd96b36f9","Type":"ContainerStarted","Data":"b1d306765f9f81233cbf4bcf9a434dad234b4f9bdf608b106738df318d28ed43"} Dec 03 13:53:41 crc kubenswrapper[4677]: I1203 13:53:41.620320 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:41 crc kubenswrapper[4677]: I1203 13:53:41.624682 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" Dec 03 13:53:41 crc kubenswrapper[4677]: I1203 13:53:41.637096 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" podStartSLOduration=3.637077216 podStartE2EDuration="3.637077216s" podCreationTimestamp="2025-12-03 13:53:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:41.634673138 +0000 UTC m=+412.381005613" watchObservedRunningTime="2025-12-03 13:53:41.637077216 +0000 UTC m=+412.383409671" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.055783 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvzwn"] Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.056544 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pvzwn" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="registry-server" containerID="cri-o://b0b995bfb18ac4a67f4d65b15cce1ff133818f8fb25278d8d04d2b0cf1c801de" gracePeriod=30 Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.068629 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2x9kl"] Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.068918 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2x9kl" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="registry-server" containerID="cri-o://208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a" gracePeriod=30 Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.075733 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lqjhf"] Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.075939 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" containerID="cri-o://d7a4e89157b13bb7b09d57dbfd96091406ea13a196ab64263f00b6ec6f2de879" gracePeriod=30 Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.082869 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbdns"] Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.083186 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vbdns" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="registry-server" containerID="cri-o://3c205eea1e80fbb53e0d6a7cf27e6155be57a527c41b414b37c01dfe79a4e113" gracePeriod=30 Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.092029 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-th96p"] Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.092452 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-th96p" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="registry-server" containerID="cri-o://f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a" gracePeriod=30 Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.104738 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfhnk"] Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.105581 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.124486 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfhnk"] Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.220113 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.220159 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.220179 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-295ls\" (UniqueName: \"kubernetes.io/projected/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-kube-api-access-295ls\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.321235 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.321276 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.321301 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-295ls\" (UniqueName: \"kubernetes.io/projected/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-kube-api-access-295ls\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.324064 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.329569 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.349195 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-295ls\" (UniqueName: \"kubernetes.io/projected/ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6-kube-api-access-295ls\") pod \"marketplace-operator-79b997595-tfhnk\" (UID: \"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.425814 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:49 crc kubenswrapper[4677]: I1203 13:53:49.874917 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfhnk"] Dec 03 13:53:49 crc kubenswrapper[4677]: E1203 13:53:49.898766 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a is running failed: container process not found" containerID="f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 13:53:49 crc kubenswrapper[4677]: E1203 13:53:49.899362 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a is running failed: container process not found" containerID="f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 13:53:49 crc kubenswrapper[4677]: E1203 13:53:49.899802 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a is running failed: container process not found" containerID="f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 13:53:49 crc kubenswrapper[4677]: E1203 13:53:49.899847 4677 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-th96p" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="registry-server" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.011220 4677 scope.go:117] "RemoveContainer" containerID="695fd537983457c862a8dc03a8f34267c35c885e3cec736e1877915adbfa5997" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.092533 4677 scope.go:117] "RemoveContainer" containerID="6993d2a824fbfa073c608eca74c7a6a5b096e436b24f0c4c621a0b9744121e4b" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.109284 4677 scope.go:117] "RemoveContainer" containerID="890bf327affe6f21744d8f9706275b536d709e6f66ed4cf946abdaef56bba2e3" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.172469 4677 scope.go:117] "RemoveContainer" containerID="4ae34073464a701a576513e9537fef0cc26e92e7a0d11af9b25613cba18d7e04" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.220109 4677 scope.go:117] "RemoveContainer" containerID="edb9c94b5ed6e6a7fc4d81a1b0bd5214c73d3e2a0c72c09737704859e55f557c" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.233329 4677 scope.go:117] "RemoveContainer" containerID="a714dab36b3a0a60b1e415699dd655ba465ffcb47817ac9bd2d1a38bddb00df3" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.676576 4677 generic.go:334] "Generic (PLEG): container finished" podID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerID="f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a" exitCode=0 Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.676671 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th96p" event={"ID":"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87","Type":"ContainerDied","Data":"f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a"} Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.678927 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" event={"ID":"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6","Type":"ContainerStarted","Data":"0aa2f033be5d886e9eebd1401ff9f5a33af6e08952050edbb8e0e6de76bf898b"} Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.682187 4677 generic.go:334] "Generic (PLEG): container finished" podID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerID="d7a4e89157b13bb7b09d57dbfd96091406ea13a196ab64263f00b6ec6f2de879" exitCode=0 Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.682335 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" event={"ID":"9dc01e2f-388f-41a2-a139-5d251ad6cda8","Type":"ContainerDied","Data":"d7a4e89157b13bb7b09d57dbfd96091406ea13a196ab64263f00b6ec6f2de879"} Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.682405 4677 scope.go:117] "RemoveContainer" containerID="9370c5745888934d7eaa524156e7b23861ec199211dc3d49521f028841ea3f51" Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.691968 4677 generic.go:334] "Generic (PLEG): container finished" podID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerID="3c205eea1e80fbb53e0d6a7cf27e6155be57a527c41b414b37c01dfe79a4e113" exitCode=0 Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.691994 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbdns" event={"ID":"dab888c4-d0b5-47b5-b0b9-176672bde399","Type":"ContainerDied","Data":"3c205eea1e80fbb53e0d6a7cf27e6155be57a527c41b414b37c01dfe79a4e113"} Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.695573 4677 generic.go:334] "Generic (PLEG): container finished" podID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerID="208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a" exitCode=0 Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.695607 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2x9kl" event={"ID":"4deaa535-2caf-4758-8470-19e0d3d2d6dd","Type":"ContainerDied","Data":"208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a"} Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.698334 4677 generic.go:334] "Generic (PLEG): container finished" podID="73e747c5-106b-479c-9264-558bf7783d87" containerID="b0b995bfb18ac4a67f4d65b15cce1ff133818f8fb25278d8d04d2b0cf1c801de" exitCode=0 Dec 03 13:53:50 crc kubenswrapper[4677]: I1203 13:53:50.698394 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvzwn" event={"ID":"73e747c5-106b-479c-9264-558bf7783d87","Type":"ContainerDied","Data":"b0b995bfb18ac4a67f4d65b15cce1ff133818f8fb25278d8d04d2b0cf1c801de"} Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.133177 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.246779 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m67kw\" (UniqueName: \"kubernetes.io/projected/dab888c4-d0b5-47b5-b0b9-176672bde399-kube-api-access-m67kw\") pod \"dab888c4-d0b5-47b5-b0b9-176672bde399\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.246979 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-catalog-content\") pod \"dab888c4-d0b5-47b5-b0b9-176672bde399\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.247050 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-utilities\") pod \"dab888c4-d0b5-47b5-b0b9-176672bde399\" (UID: \"dab888c4-d0b5-47b5-b0b9-176672bde399\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.248278 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-utilities" (OuterVolumeSpecName: "utilities") pod "dab888c4-d0b5-47b5-b0b9-176672bde399" (UID: "dab888c4-d0b5-47b5-b0b9-176672bde399"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.258820 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab888c4-d0b5-47b5-b0b9-176672bde399-kube-api-access-m67kw" (OuterVolumeSpecName: "kube-api-access-m67kw") pod "dab888c4-d0b5-47b5-b0b9-176672bde399" (UID: "dab888c4-d0b5-47b5-b0b9-176672bde399"). InnerVolumeSpecName "kube-api-access-m67kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.273474 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dab888c4-d0b5-47b5-b0b9-176672bde399" (UID: "dab888c4-d0b5-47b5-b0b9-176672bde399"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.348961 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.349003 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab888c4-d0b5-47b5-b0b9-176672bde399-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.349018 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m67kw\" (UniqueName: \"kubernetes.io/projected/dab888c4-d0b5-47b5-b0b9-176672bde399-kube-api-access-m67kw\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.409001 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.416030 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.423004 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.430511 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.450640 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-utilities\") pod \"73e747c5-106b-479c-9264-558bf7783d87\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.450757 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-catalog-content\") pod \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.450793 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-catalog-content\") pod \"73e747c5-106b-479c-9264-558bf7783d87\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.450820 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdmbn\" (UniqueName: \"kubernetes.io/projected/4deaa535-2caf-4758-8470-19e0d3d2d6dd-kube-api-access-cdmbn\") pod \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.450845 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwgww\" (UniqueName: \"kubernetes.io/projected/73e747c5-106b-479c-9264-558bf7783d87-kube-api-access-xwgww\") pod \"73e747c5-106b-479c-9264-558bf7783d87\" (UID: \"73e747c5-106b-479c-9264-558bf7783d87\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.450917 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-utilities\") pod \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\" (UID: \"4deaa535-2caf-4758-8470-19e0d3d2d6dd\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.453380 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-utilities" (OuterVolumeSpecName: "utilities") pod "4deaa535-2caf-4758-8470-19e0d3d2d6dd" (UID: "4deaa535-2caf-4758-8470-19e0d3d2d6dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.455399 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-utilities" (OuterVolumeSpecName: "utilities") pod "73e747c5-106b-479c-9264-558bf7783d87" (UID: "73e747c5-106b-479c-9264-558bf7783d87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.469156 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4deaa535-2caf-4758-8470-19e0d3d2d6dd-kube-api-access-cdmbn" (OuterVolumeSpecName: "kube-api-access-cdmbn") pod "4deaa535-2caf-4758-8470-19e0d3d2d6dd" (UID: "4deaa535-2caf-4758-8470-19e0d3d2d6dd"). InnerVolumeSpecName "kube-api-access-cdmbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.477179 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e747c5-106b-479c-9264-558bf7783d87-kube-api-access-xwgww" (OuterVolumeSpecName: "kube-api-access-xwgww") pod "73e747c5-106b-479c-9264-558bf7783d87" (UID: "73e747c5-106b-479c-9264-558bf7783d87"). InnerVolumeSpecName "kube-api-access-xwgww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.551632 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2nbl\" (UniqueName: \"kubernetes.io/projected/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-kube-api-access-c2nbl\") pod \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.551696 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-catalog-content\") pod \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.551727 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-operator-metrics\") pod \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.551766 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4sgb\" (UniqueName: \"kubernetes.io/projected/9dc01e2f-388f-41a2-a139-5d251ad6cda8-kube-api-access-z4sgb\") pod \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.551788 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-utilities\") pod \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\" (UID: \"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.551813 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-trusted-ca\") pod \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\" (UID: \"9dc01e2f-388f-41a2-a139-5d251ad6cda8\") " Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.552054 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdmbn\" (UniqueName: \"kubernetes.io/projected/4deaa535-2caf-4758-8470-19e0d3d2d6dd-kube-api-access-cdmbn\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.552067 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwgww\" (UniqueName: \"kubernetes.io/projected/73e747c5-106b-479c-9264-558bf7783d87-kube-api-access-xwgww\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.552076 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.552084 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.552657 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9dc01e2f-388f-41a2-a139-5d251ad6cda8" (UID: "9dc01e2f-388f-41a2-a139-5d251ad6cda8"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.559513 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-utilities" (OuterVolumeSpecName: "utilities") pod "fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" (UID: "fde6c082-8a6f-4e3b-bf24-68be4bfdcf87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.566540 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dc01e2f-388f-41a2-a139-5d251ad6cda8-kube-api-access-z4sgb" (OuterVolumeSpecName: "kube-api-access-z4sgb") pod "9dc01e2f-388f-41a2-a139-5d251ad6cda8" (UID: "9dc01e2f-388f-41a2-a139-5d251ad6cda8"). InnerVolumeSpecName "kube-api-access-z4sgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.568170 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-kube-api-access-c2nbl" (OuterVolumeSpecName: "kube-api-access-c2nbl") pod "fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" (UID: "fde6c082-8a6f-4e3b-bf24-68be4bfdcf87"). InnerVolumeSpecName "kube-api-access-c2nbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.574638 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9dc01e2f-388f-41a2-a139-5d251ad6cda8" (UID: "9dc01e2f-388f-41a2-a139-5d251ad6cda8"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.595303 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73e747c5-106b-479c-9264-558bf7783d87" (UID: "73e747c5-106b-479c-9264-558bf7783d87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.595996 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4deaa535-2caf-4758-8470-19e0d3d2d6dd" (UID: "4deaa535-2caf-4758-8470-19e0d3d2d6dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.653756 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4deaa535-2caf-4758-8470-19e0d3d2d6dd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.654235 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e747c5-106b-479c-9264-558bf7783d87-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.654252 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2nbl\" (UniqueName: \"kubernetes.io/projected/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-kube-api-access-c2nbl\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.654269 4677 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.654282 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4sgb\" (UniqueName: \"kubernetes.io/projected/9dc01e2f-388f-41a2-a139-5d251ad6cda8-kube-api-access-z4sgb\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.654318 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.654333 4677 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9dc01e2f-388f-41a2-a139-5d251ad6cda8-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.676326 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" (UID: "fde6c082-8a6f-4e3b-bf24-68be4bfdcf87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.706482 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" event={"ID":"ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6","Type":"ContainerStarted","Data":"be8c0c3aa244c2cc4c6c50fa25c487785355df10fee730aba77d79be8b128674"} Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.710541 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvzwn" event={"ID":"73e747c5-106b-479c-9264-558bf7783d87","Type":"ContainerDied","Data":"3ec2abab01f797b0643275babc158ef428eabe3ca5a9e83fb7498954f3f7ee1c"} Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.710633 4677 scope.go:117] "RemoveContainer" containerID="b0b995bfb18ac4a67f4d65b15cce1ff133818f8fb25278d8d04d2b0cf1c801de" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.710554 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvzwn" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.714536 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbdns" event={"ID":"dab888c4-d0b5-47b5-b0b9-176672bde399","Type":"ContainerDied","Data":"707b724a9411b864293ca5acacf8cd87e7e12207ad9a2958ae8c249411319cdf"} Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.714662 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbdns" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.717578 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2x9kl" event={"ID":"4deaa535-2caf-4758-8470-19e0d3d2d6dd","Type":"ContainerDied","Data":"c05ed6eca5c209971450d097770c8d666e150f618c2b489702ca452acd99ad72"} Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.717703 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2x9kl" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.725139 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" event={"ID":"9dc01e2f-388f-41a2-a139-5d251ad6cda8","Type":"ContainerDied","Data":"bdab2684993e917569777b4e9fe89e67087b0f8c42b9660818543cd9631c95e0"} Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.726059 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lqjhf" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.729694 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th96p" event={"ID":"fde6c082-8a6f-4e3b-bf24-68be4bfdcf87","Type":"ContainerDied","Data":"a9da1fa851c20e7fb00885ac62be74ad2da3e290fd5dfec888449b453e963a62"} Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.729815 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th96p" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.736204 4677 scope.go:117] "RemoveContainer" containerID="84e53a5a21cd43fe8f6a743c206621b0e9765a1d7ea2bc55d96d3327b59ebf08" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.743379 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" podStartSLOduration=2.743349286 podStartE2EDuration="2.743349286s" podCreationTimestamp="2025-12-03 13:53:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:51.722732301 +0000 UTC m=+422.469064776" watchObservedRunningTime="2025-12-03 13:53:51.743349286 +0000 UTC m=+422.489681761" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.759774 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.756128 4677 scope.go:117] "RemoveContainer" containerID="9262255b00bdb2b4906c3c557f398012d0b88a4f46059c5cc04cbed27c988b36" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.781989 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvzwn"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.806080 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pvzwn"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.812422 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbdns"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.814380 4677 scope.go:117] "RemoveContainer" containerID="3c205eea1e80fbb53e0d6a7cf27e6155be57a527c41b414b37c01dfe79a4e113" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.816288 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbdns"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.820371 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-th96p"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.823225 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-th96p"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.826452 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lqjhf"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.838126 4677 scope.go:117] "RemoveContainer" containerID="d39f678c854e6e145ee1f2ae7514cf96d300db7c527afa8ed27af30bb4a07c1b" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.841854 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lqjhf"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.845725 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2x9kl"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.849423 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2x9kl"] Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.851606 4677 scope.go:117] "RemoveContainer" containerID="0d5e42547fe6a3d30f7695c616007c8c59b36c66444d011114f13c0058fd77d7" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.865170 4677 scope.go:117] "RemoveContainer" containerID="208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.881591 4677 scope.go:117] "RemoveContainer" containerID="15675fd847f4cead6f2819d806eb6587f4a4055bfaed31ada9f67f90279c1edb" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.895847 4677 scope.go:117] "RemoveContainer" containerID="ae21a2713ae1e81f38be1fee15332a7ba25790d9b474a670714307062d256486" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.908618 4677 scope.go:117] "RemoveContainer" containerID="d7a4e89157b13bb7b09d57dbfd96091406ea13a196ab64263f00b6ec6f2de879" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.923672 4677 scope.go:117] "RemoveContainer" containerID="f29a522b8eb0fa9099b9f7307e6fa53c51198f5cbc9cef1ae570ad926a03988a" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.939371 4677 scope.go:117] "RemoveContainer" containerID="1a9bca2aae97b40230bfa8ca39db60a2d916a15de70d02b0414411ce20def3b2" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.954132 4677 scope.go:117] "RemoveContainer" containerID="ab96a988e526fbbcdc2854cdbee810777a6c8a992507911101a5f6ac0a1d034c" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.983848 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" path="/var/lib/kubelet/pods/4deaa535-2caf-4758-8470-19e0d3d2d6dd/volumes" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.984543 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e747c5-106b-479c-9264-558bf7783d87" path="/var/lib/kubelet/pods/73e747c5-106b-479c-9264-558bf7783d87/volumes" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.985205 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" path="/var/lib/kubelet/pods/9dc01e2f-388f-41a2-a139-5d251ad6cda8/volumes" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.986215 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" path="/var/lib/kubelet/pods/dab888c4-d0b5-47b5-b0b9-176672bde399/volumes" Dec 03 13:53:51 crc kubenswrapper[4677]: I1203 13:53:51.986827 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" path="/var/lib/kubelet/pods/fde6c082-8a6f-4e3b-bf24-68be4bfdcf87/volumes" Dec 03 13:53:52 crc kubenswrapper[4677]: I1203 13:53:52.739163 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:52 crc kubenswrapper[4677]: I1203 13:53:52.742936 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tfhnk" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.288534 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-68vg8"] Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289089 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289127 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289148 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289163 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289181 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289198 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289224 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289242 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289261 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289275 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289297 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289309 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289330 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289343 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289364 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289376 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289460 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289473 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289491 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289503 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289520 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289532 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289546 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289557 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="extract-content" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289576 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289587 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: E1203 13:53:53.289663 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289676 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="extract-utilities" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289895 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="4deaa535-2caf-4758-8470-19e0d3d2d6dd" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289921 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.289939 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde6c082-8a6f-4e3b-bf24-68be4bfdcf87" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.290001 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e747c5-106b-479c-9264-558bf7783d87" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.290019 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab888c4-d0b5-47b5-b0b9-176672bde399" containerName="registry-server" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.290565 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dc01e2f-388f-41a2-a139-5d251ad6cda8" containerName="marketplace-operator" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.291787 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.294516 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.297156 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68vg8"] Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.384340 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv2jj\" (UniqueName: \"kubernetes.io/projected/361c4f16-5752-4a54-bae4-033bc1425b51-kube-api-access-dv2jj\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.384429 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/361c4f16-5752-4a54-bae4-033bc1425b51-utilities\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.384571 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/361c4f16-5752-4a54-bae4-033bc1425b51-catalog-content\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.484593 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hm4kc"] Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.485657 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/361c4f16-5752-4a54-bae4-033bc1425b51-catalog-content\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.485693 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.485726 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv2jj\" (UniqueName: \"kubernetes.io/projected/361c4f16-5752-4a54-bae4-033bc1425b51-kube-api-access-dv2jj\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.485780 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/361c4f16-5752-4a54-bae4-033bc1425b51-utilities\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.486241 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/361c4f16-5752-4a54-bae4-033bc1425b51-catalog-content\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.486258 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/361c4f16-5752-4a54-bae4-033bc1425b51-utilities\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.488068 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.495126 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hm4kc"] Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.512835 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv2jj\" (UniqueName: \"kubernetes.io/projected/361c4f16-5752-4a54-bae4-033bc1425b51-kube-api-access-dv2jj\") pod \"community-operators-68vg8\" (UID: \"361c4f16-5752-4a54-bae4-033bc1425b51\") " pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.587514 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b895f39-0aeb-4a61-857d-9d9e91b103f3-utilities\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.587604 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b895f39-0aeb-4a61-857d-9d9e91b103f3-catalog-content\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.587636 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhwmj\" (UniqueName: \"kubernetes.io/projected/5b895f39-0aeb-4a61-857d-9d9e91b103f3-kube-api-access-xhwmj\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.605590 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.688918 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b895f39-0aeb-4a61-857d-9d9e91b103f3-utilities\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.689028 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b895f39-0aeb-4a61-857d-9d9e91b103f3-catalog-content\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.689059 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhwmj\" (UniqueName: \"kubernetes.io/projected/5b895f39-0aeb-4a61-857d-9d9e91b103f3-kube-api-access-xhwmj\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.690047 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b895f39-0aeb-4a61-857d-9d9e91b103f3-utilities\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.690345 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b895f39-0aeb-4a61-857d-9d9e91b103f3-catalog-content\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.712499 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhwmj\" (UniqueName: \"kubernetes.io/projected/5b895f39-0aeb-4a61-857d-9d9e91b103f3-kube-api-access-xhwmj\") pod \"certified-operators-hm4kc\" (UID: \"5b895f39-0aeb-4a61-857d-9d9e91b103f3\") " pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:53 crc kubenswrapper[4677]: I1203 13:53:53.807842 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.005713 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68vg8"] Dec 03 13:53:54 crc kubenswrapper[4677]: W1203 13:53:54.013513 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod361c4f16_5752_4a54_bae4_033bc1425b51.slice/crio-137088a98c49019a80d676e0076a0c57f95139a736c13e8d31ae852fa40ecac7 WatchSource:0}: Error finding container 137088a98c49019a80d676e0076a0c57f95139a736c13e8d31ae852fa40ecac7: Status 404 returned error can't find the container with id 137088a98c49019a80d676e0076a0c57f95139a736c13e8d31ae852fa40ecac7 Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.190449 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hm4kc"] Dec 03 13:53:54 crc kubenswrapper[4677]: E1203 13:53:54.318171 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4deaa535_2caf_4758_8470_19e0d3d2d6dd.slice/crio-conmon-208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.752860 4677 generic.go:334] "Generic (PLEG): container finished" podID="361c4f16-5752-4a54-bae4-033bc1425b51" containerID="805661437bc3286b0dfffbb6ea5aabe7547bfca1b6475bf610412d150103d563" exitCode=0 Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.752939 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68vg8" event={"ID":"361c4f16-5752-4a54-bae4-033bc1425b51","Type":"ContainerDied","Data":"805661437bc3286b0dfffbb6ea5aabe7547bfca1b6475bf610412d150103d563"} Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.754001 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68vg8" event={"ID":"361c4f16-5752-4a54-bae4-033bc1425b51","Type":"ContainerStarted","Data":"137088a98c49019a80d676e0076a0c57f95139a736c13e8d31ae852fa40ecac7"} Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.757680 4677 generic.go:334] "Generic (PLEG): container finished" podID="5b895f39-0aeb-4a61-857d-9d9e91b103f3" containerID="6ef55db71a4532afe9b594bf6ad48d3c2826223e15b81278dff42013d2b9cac4" exitCode=0 Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.757761 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm4kc" event={"ID":"5b895f39-0aeb-4a61-857d-9d9e91b103f3","Type":"ContainerDied","Data":"6ef55db71a4532afe9b594bf6ad48d3c2826223e15b81278dff42013d2b9cac4"} Dec 03 13:53:54 crc kubenswrapper[4677]: I1203 13:53:54.757820 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm4kc" event={"ID":"5b895f39-0aeb-4a61-857d-9d9e91b103f3","Type":"ContainerStarted","Data":"f168e2b5240dff1d5fad22ea98bf9d6996d9e5e66c6e7eee8fec96f15d6ffc3e"} Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.681445 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7n5sl"] Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.683038 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.697521 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.700719 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7n5sl"] Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.706280 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-x9htr" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.715170 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-utilities\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.715205 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-catalog-content\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.715239 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxq7t\" (UniqueName: \"kubernetes.io/projected/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-kube-api-access-dxq7t\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.762712 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hct2b"] Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.817330 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-utilities\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.817373 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-catalog-content\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.817816 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-catalog-content\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.817744 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-utilities\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.817810 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxq7t\" (UniqueName: \"kubernetes.io/projected/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-kube-api-access-dxq7t\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.844969 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxq7t\" (UniqueName: \"kubernetes.io/projected/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-kube-api-access-dxq7t\") pod \"redhat-operators-7n5sl\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.878280 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s4dpj"] Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.879482 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.881818 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.893687 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4dpj"] Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.919029 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z57tq\" (UniqueName: \"kubernetes.io/projected/d1c06aff-5cb1-4653-a4ed-bfddfab19938-kube-api-access-z57tq\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.919091 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c06aff-5cb1-4653-a4ed-bfddfab19938-catalog-content\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:55 crc kubenswrapper[4677]: I1203 13:53:55.919380 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c06aff-5cb1-4653-a4ed-bfddfab19938-utilities\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.018291 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.020422 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z57tq\" (UniqueName: \"kubernetes.io/projected/d1c06aff-5cb1-4653-a4ed-bfddfab19938-kube-api-access-z57tq\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.020482 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c06aff-5cb1-4653-a4ed-bfddfab19938-catalog-content\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.020553 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c06aff-5cb1-4653-a4ed-bfddfab19938-utilities\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.021376 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c06aff-5cb1-4653-a4ed-bfddfab19938-utilities\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.021537 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c06aff-5cb1-4653-a4ed-bfddfab19938-catalog-content\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.047798 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z57tq\" (UniqueName: \"kubernetes.io/projected/d1c06aff-5cb1-4653-a4ed-bfddfab19938-kube-api-access-z57tq\") pod \"redhat-marketplace-s4dpj\" (UID: \"d1c06aff-5cb1-4653-a4ed-bfddfab19938\") " pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.205934 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.448942 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7n5sl"] Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.588555 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4dpj"] Dec 03 13:53:56 crc kubenswrapper[4677]: W1203 13:53:56.596243 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1c06aff_5cb1_4653_a4ed_bfddfab19938.slice/crio-ddd115a2a8e77857fa3c61c166e7531123773c681587fa9b0921a1b134e1419f WatchSource:0}: Error finding container ddd115a2a8e77857fa3c61c166e7531123773c681587fa9b0921a1b134e1419f: Status 404 returned error can't find the container with id ddd115a2a8e77857fa3c61c166e7531123773c681587fa9b0921a1b134e1419f Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.772587 4677 generic.go:334] "Generic (PLEG): container finished" podID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerID="954724ac90d374ece8ed221777465e7e8782cbae90e4ae910b2838e0bcfd12bc" exitCode=0 Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.772665 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7n5sl" event={"ID":"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c","Type":"ContainerDied","Data":"954724ac90d374ece8ed221777465e7e8782cbae90e4ae910b2838e0bcfd12bc"} Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.772919 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7n5sl" event={"ID":"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c","Type":"ContainerStarted","Data":"2868e966fae95a263f7e580a4f35a4e3e5cd080b0390d80c9124db1ea39bd1fc"} Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.775409 4677 generic.go:334] "Generic (PLEG): container finished" podID="d1c06aff-5cb1-4653-a4ed-bfddfab19938" containerID="e446636a06f2b1d07397b3d71cf8f7770df46ef9d3cbe9994bd6fb8c5e408009" exitCode=0 Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.775435 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4dpj" event={"ID":"d1c06aff-5cb1-4653-a4ed-bfddfab19938","Type":"ContainerDied","Data":"e446636a06f2b1d07397b3d71cf8f7770df46ef9d3cbe9994bd6fb8c5e408009"} Dec 03 13:53:56 crc kubenswrapper[4677]: I1203 13:53:56.775453 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4dpj" event={"ID":"d1c06aff-5cb1-4653-a4ed-bfddfab19938","Type":"ContainerStarted","Data":"ddd115a2a8e77857fa3c61c166e7531123773c681587fa9b0921a1b134e1419f"} Dec 03 13:53:57 crc kubenswrapper[4677]: I1203 13:53:57.800561 4677 generic.go:334] "Generic (PLEG): container finished" podID="361c4f16-5752-4a54-bae4-033bc1425b51" containerID="9b6c43071736676bf7314cfe8de5f9f39ce3a6906252793e4b020ddeb87da62f" exitCode=0 Dec 03 13:53:57 crc kubenswrapper[4677]: I1203 13:53:57.800637 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68vg8" event={"ID":"361c4f16-5752-4a54-bae4-033bc1425b51","Type":"ContainerDied","Data":"9b6c43071736676bf7314cfe8de5f9f39ce3a6906252793e4b020ddeb87da62f"} Dec 03 13:53:57 crc kubenswrapper[4677]: I1203 13:53:57.804119 4677 generic.go:334] "Generic (PLEG): container finished" podID="5b895f39-0aeb-4a61-857d-9d9e91b103f3" containerID="2afaa0c86bc6933d0f7498864a872e87e2dc92276fa36a89a71b5bd15b9d835c" exitCode=0 Dec 03 13:53:57 crc kubenswrapper[4677]: I1203 13:53:57.804146 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm4kc" event={"ID":"5b895f39-0aeb-4a61-857d-9d9e91b103f3","Type":"ContainerDied","Data":"2afaa0c86bc6933d0f7498864a872e87e2dc92276fa36a89a71b5bd15b9d835c"} Dec 03 13:53:58 crc kubenswrapper[4677]: I1203 13:53:58.815156 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm4kc" event={"ID":"5b895f39-0aeb-4a61-857d-9d9e91b103f3","Type":"ContainerStarted","Data":"d7c06e82008a5d96f00aa01de90bd84bd326d70cce879c92938b528fb1bf81aa"} Dec 03 13:53:58 crc kubenswrapper[4677]: I1203 13:53:58.817324 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7n5sl" event={"ID":"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c","Type":"ContainerStarted","Data":"0c27748c1cb7898c4f024a01a0138c7f8d63ed2266359d8dc18ac3bfd0e574de"} Dec 03 13:53:58 crc kubenswrapper[4677]: I1203 13:53:58.820732 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68vg8" event={"ID":"361c4f16-5752-4a54-bae4-033bc1425b51","Type":"ContainerStarted","Data":"a107a3ca2ba1c14541e98bac6fff95d1b814457763499850b167e7b65ef1155f"} Dec 03 13:53:58 crc kubenswrapper[4677]: I1203 13:53:58.824465 4677 generic.go:334] "Generic (PLEG): container finished" podID="d1c06aff-5cb1-4653-a4ed-bfddfab19938" containerID="3d0280ed5587b1d55669cc866be029c18471fac3693c6d5edf5d9b730c0d8bc3" exitCode=0 Dec 03 13:53:58 crc kubenswrapper[4677]: I1203 13:53:58.824512 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4dpj" event={"ID":"d1c06aff-5cb1-4653-a4ed-bfddfab19938","Type":"ContainerDied","Data":"3d0280ed5587b1d55669cc866be029c18471fac3693c6d5edf5d9b730c0d8bc3"} Dec 03 13:53:58 crc kubenswrapper[4677]: I1203 13:53:58.844344 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hm4kc" podStartSLOduration=2.047692889 podStartE2EDuration="5.844254818s" podCreationTimestamp="2025-12-03 13:53:53 +0000 UTC" firstStartedPulling="2025-12-03 13:53:54.760128243 +0000 UTC m=+425.506460698" lastFinishedPulling="2025-12-03 13:53:58.556690172 +0000 UTC m=+429.303022627" observedRunningTime="2025-12-03 13:53:58.838324783 +0000 UTC m=+429.584657238" watchObservedRunningTime="2025-12-03 13:53:58.844254818 +0000 UTC m=+429.590587293" Dec 03 13:53:58 crc kubenswrapper[4677]: I1203 13:53:58.879397 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-68vg8" podStartSLOduration=2.147676098 podStartE2EDuration="5.879373748s" podCreationTimestamp="2025-12-03 13:53:53 +0000 UTC" firstStartedPulling="2025-12-03 13:53:54.755359699 +0000 UTC m=+425.501692164" lastFinishedPulling="2025-12-03 13:53:58.487057359 +0000 UTC m=+429.233389814" observedRunningTime="2025-12-03 13:53:58.875995844 +0000 UTC m=+429.622328319" watchObservedRunningTime="2025-12-03 13:53:58.879373748 +0000 UTC m=+429.625706203" Dec 03 13:53:59 crc kubenswrapper[4677]: I1203 13:53:59.832288 4677 generic.go:334] "Generic (PLEG): container finished" podID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerID="0c27748c1cb7898c4f024a01a0138c7f8d63ed2266359d8dc18ac3bfd0e574de" exitCode=0 Dec 03 13:53:59 crc kubenswrapper[4677]: I1203 13:53:59.832430 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7n5sl" event={"ID":"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c","Type":"ContainerDied","Data":"0c27748c1cb7898c4f024a01a0138c7f8d63ed2266359d8dc18ac3bfd0e574de"} Dec 03 13:54:00 crc kubenswrapper[4677]: I1203 13:54:00.844625 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4dpj" event={"ID":"d1c06aff-5cb1-4653-a4ed-bfddfab19938","Type":"ContainerStarted","Data":"2b0a461d6286368a041721e14c8861f6a45cc3156429cf0c30eac16a43e30de4"} Dec 03 13:54:00 crc kubenswrapper[4677]: I1203 13:54:00.862150 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s4dpj" podStartSLOduration=3.31009428 podStartE2EDuration="5.862135896s" podCreationTimestamp="2025-12-03 13:53:55 +0000 UTC" firstStartedPulling="2025-12-03 13:53:56.777877277 +0000 UTC m=+427.524209732" lastFinishedPulling="2025-12-03 13:53:59.329918893 +0000 UTC m=+430.076251348" observedRunningTime="2025-12-03 13:54:00.861105017 +0000 UTC m=+431.607437482" watchObservedRunningTime="2025-12-03 13:54:00.862135896 +0000 UTC m=+431.608468351" Dec 03 13:54:01 crc kubenswrapper[4677]: I1203 13:54:01.854314 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7n5sl" event={"ID":"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c","Type":"ContainerStarted","Data":"b1f8053ecee7e87b26fbb81dea6dc971e944c67374d15131201fb65e2ccf77b6"} Dec 03 13:54:01 crc kubenswrapper[4677]: I1203 13:54:01.872665 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7n5sl" podStartSLOduration=2.272501041 podStartE2EDuration="6.872648869s" podCreationTimestamp="2025-12-03 13:53:55 +0000 UTC" firstStartedPulling="2025-12-03 13:53:56.792854174 +0000 UTC m=+427.539186629" lastFinishedPulling="2025-12-03 13:54:01.393001992 +0000 UTC m=+432.139334457" observedRunningTime="2025-12-03 13:54:01.872258648 +0000 UTC m=+432.618591113" watchObservedRunningTime="2025-12-03 13:54:01.872648869 +0000 UTC m=+432.618981324" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.605747 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.611863 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.655024 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.810510 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.810681 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.854186 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.898052 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-68vg8" Dec 03 13:54:03 crc kubenswrapper[4677]: I1203 13:54:03.903694 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hm4kc" Dec 03 13:54:04 crc kubenswrapper[4677]: E1203 13:54:04.451189 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4deaa535_2caf_4758_8470_19e0d3d2d6dd.slice/crio-conmon-208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:54:06 crc kubenswrapper[4677]: I1203 13:54:06.019345 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:54:06 crc kubenswrapper[4677]: I1203 13:54:06.019681 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:54:06 crc kubenswrapper[4677]: I1203 13:54:06.207060 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:54:06 crc kubenswrapper[4677]: I1203 13:54:06.207445 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:54:06 crc kubenswrapper[4677]: I1203 13:54:06.254563 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:54:06 crc kubenswrapper[4677]: I1203 13:54:06.914439 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s4dpj" Dec 03 13:54:07 crc kubenswrapper[4677]: I1203 13:54:07.056438 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7n5sl" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="registry-server" probeResult="failure" output=< Dec 03 13:54:07 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 13:54:07 crc kubenswrapper[4677]: > Dec 03 13:54:08 crc kubenswrapper[4677]: I1203 13:54:08.436792 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:54:08 crc kubenswrapper[4677]: I1203 13:54:08.436853 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:54:08 crc kubenswrapper[4677]: I1203 13:54:08.436903 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:54:08 crc kubenswrapper[4677]: I1203 13:54:08.437510 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58bddf0714a1025ce6f4421667e6b66749f342b2705111136662a9104f1b7753"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:54:08 crc kubenswrapper[4677]: I1203 13:54:08.437561 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://58bddf0714a1025ce6f4421667e6b66749f342b2705111136662a9104f1b7753" gracePeriod=600 Dec 03 13:54:14 crc kubenswrapper[4677]: E1203 13:54:14.568826 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4deaa535_2caf_4758_8470_19e0d3d2d6dd.slice/crio-conmon-208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:54:14 crc kubenswrapper[4677]: I1203 13:54:14.745714 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="58bddf0714a1025ce6f4421667e6b66749f342b2705111136662a9104f1b7753" exitCode=0 Dec 03 13:54:14 crc kubenswrapper[4677]: I1203 13:54:14.745781 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"58bddf0714a1025ce6f4421667e6b66749f342b2705111136662a9104f1b7753"} Dec 03 13:54:14 crc kubenswrapper[4677]: I1203 13:54:14.745845 4677 scope.go:117] "RemoveContainer" containerID="c72ad30399007f767efb6c2f484ad3f19f1e4344cdaefb75ff46ee6f2eb106b4" Dec 03 13:54:16 crc kubenswrapper[4677]: I1203 13:54:16.074321 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:54:16 crc kubenswrapper[4677]: I1203 13:54:16.121272 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 13:54:16 crc kubenswrapper[4677]: I1203 13:54:16.759619 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"25cc84bb4570658cc4097ad0606cd34b6dcadcbff9d0daecfbed41321f21143e"} Dec 03 13:54:20 crc kubenswrapper[4677]: I1203 13:54:20.800869 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" podUID="0963f4ad-95a1-4602-ab12-3073f3db0581" containerName="registry" containerID="cri-o://e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1" gracePeriod=30 Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.259108 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.437108 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwzpd\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-kube-api-access-kwzpd\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.437187 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-bound-sa-token\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.437240 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0963f4ad-95a1-4602-ab12-3073f3db0581-ca-trust-extracted\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.437269 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-tls\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.438123 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.438152 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-trusted-ca\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.438173 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0963f4ad-95a1-4602-ab12-3073f3db0581-installation-pull-secrets\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.438202 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-certificates\") pod \"0963f4ad-95a1-4602-ab12-3073f3db0581\" (UID: \"0963f4ad-95a1-4602-ab12-3073f3db0581\") " Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.438997 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.440271 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.443335 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.444076 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0963f4ad-95a1-4602-ab12-3073f3db0581-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.444475 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.445577 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-kube-api-access-kwzpd" (OuterVolumeSpecName: "kube-api-access-kwzpd") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "kube-api-access-kwzpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.448600 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.458683 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0963f4ad-95a1-4602-ab12-3073f3db0581-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "0963f4ad-95a1-4602-ab12-3073f3db0581" (UID: "0963f4ad-95a1-4602-ab12-3073f3db0581"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.539273 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.539636 4677 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0963f4ad-95a1-4602-ab12-3073f3db0581-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.539651 4677 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.539662 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwzpd\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-kube-api-access-kwzpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.539672 4677 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.539681 4677 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0963f4ad-95a1-4602-ab12-3073f3db0581-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.539689 4677 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0963f4ad-95a1-4602-ab12-3073f3db0581-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.788396 4677 generic.go:334] "Generic (PLEG): container finished" podID="0963f4ad-95a1-4602-ab12-3073f3db0581" containerID="e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1" exitCode=0 Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.788444 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" event={"ID":"0963f4ad-95a1-4602-ab12-3073f3db0581","Type":"ContainerDied","Data":"e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1"} Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.788468 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.788484 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hct2b" event={"ID":"0963f4ad-95a1-4602-ab12-3073f3db0581","Type":"ContainerDied","Data":"370351aa33c17df7adefca93a32ec225c1aca4c6dc1c19b51637f000fb0b04d1"} Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.788507 4677 scope.go:117] "RemoveContainer" containerID="e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.807173 4677 scope.go:117] "RemoveContainer" containerID="e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1" Dec 03 13:54:21 crc kubenswrapper[4677]: E1203 13:54:21.808728 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1\": container with ID starting with e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1 not found: ID does not exist" containerID="e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.808760 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1"} err="failed to get container status \"e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1\": rpc error: code = NotFound desc = could not find container \"e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1\": container with ID starting with e4cd7c446a3aef349769dba4cd130c8151c55d7288c6e4c1d9ad2d966fcbc3b1 not found: ID does not exist" Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.822126 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hct2b"] Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.827041 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hct2b"] Dec 03 13:54:21 crc kubenswrapper[4677]: I1203 13:54:21.982422 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0963f4ad-95a1-4602-ab12-3073f3db0581" path="/var/lib/kubelet/pods/0963f4ad-95a1-4602-ab12-3073f3db0581/volumes" Dec 03 13:54:24 crc kubenswrapper[4677]: E1203 13:54:24.674653 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4deaa535_2caf_4758_8470_19e0d3d2d6dd.slice/crio-conmon-208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:54:34 crc kubenswrapper[4677]: E1203 13:54:34.782092 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4deaa535_2caf_4758_8470_19e0d3d2d6dd.slice/crio-conmon-208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:54:44 crc kubenswrapper[4677]: E1203 13:54:44.908264 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4deaa535_2caf_4758_8470_19e0d3d2d6dd.slice/crio-conmon-208714f5471e90b5177c1e1c1aa0db8115cae491a48eade00aeafc382e08332a.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:56:38 crc kubenswrapper[4677]: I1203 13:56:38.437662 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:56:38 crc kubenswrapper[4677]: I1203 13:56:38.438471 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:57:08 crc kubenswrapper[4677]: I1203 13:57:08.437867 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:57:08 crc kubenswrapper[4677]: I1203 13:57:08.438442 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.437176 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.437867 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.437992 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.438923 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25cc84bb4570658cc4097ad0606cd34b6dcadcbff9d0daecfbed41321f21143e"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.439064 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://25cc84bb4570658cc4097ad0606cd34b6dcadcbff9d0daecfbed41321f21143e" gracePeriod=600 Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.776549 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="25cc84bb4570658cc4097ad0606cd34b6dcadcbff9d0daecfbed41321f21143e" exitCode=0 Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.776640 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"25cc84bb4570658cc4097ad0606cd34b6dcadcbff9d0daecfbed41321f21143e"} Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.776982 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"66093cdea49042d1d130e4c423008a09d5d1d6840fa98fa247eeb0b6e0fa1441"} Dec 03 13:57:38 crc kubenswrapper[4677]: I1203 13:57:38.777009 4677 scope.go:117] "RemoveContainer" containerID="58bddf0714a1025ce6f4421667e6b66749f342b2705111136662a9104f1b7753" Dec 03 13:59:30 crc kubenswrapper[4677]: I1203 13:59:30.024390 4677 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 13:59:38 crc kubenswrapper[4677]: I1203 13:59:38.436869 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:59:38 crc kubenswrapper[4677]: I1203 13:59:38.437460 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.168907 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z"] Dec 03 14:00:00 crc kubenswrapper[4677]: E1203 14:00:00.169708 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0963f4ad-95a1-4602-ab12-3073f3db0581" containerName="registry" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.169728 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0963f4ad-95a1-4602-ab12-3073f3db0581" containerName="registry" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.169855 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0963f4ad-95a1-4602-ab12-3073f3db0581" containerName="registry" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.170331 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.177191 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.180877 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.185047 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z"] Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.359442 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-config-volume\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.359534 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-secret-volume\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.359575 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwp7g\" (UniqueName: \"kubernetes.io/projected/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-kube-api-access-rwp7g\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.461048 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-secret-volume\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.461193 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwp7g\" (UniqueName: \"kubernetes.io/projected/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-kube-api-access-rwp7g\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.461274 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-config-volume\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.463223 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-config-volume\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.475625 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-secret-volume\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.480475 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwp7g\" (UniqueName: \"kubernetes.io/projected/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-kube-api-access-rwp7g\") pod \"collect-profiles-29412840-6887z\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.493116 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:00 crc kubenswrapper[4677]: I1203 14:00:00.724519 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z"] Dec 03 14:00:01 crc kubenswrapper[4677]: I1203 14:00:01.578780 4677 generic.go:334] "Generic (PLEG): container finished" podID="f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" containerID="5be8426032319e3c821a9e57f3a73205a4327f696e24458705383cef2a398bac" exitCode=0 Dec 03 14:00:01 crc kubenswrapper[4677]: I1203 14:00:01.578877 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" event={"ID":"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8","Type":"ContainerDied","Data":"5be8426032319e3c821a9e57f3a73205a4327f696e24458705383cef2a398bac"} Dec 03 14:00:01 crc kubenswrapper[4677]: I1203 14:00:01.579149 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" event={"ID":"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8","Type":"ContainerStarted","Data":"cec68dcf1c4cc6c1f3c74db6e714aeca110f657ca2365dca17cb2695f4460b97"} Dec 03 14:00:02 crc kubenswrapper[4677]: I1203 14:00:02.825020 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:02 crc kubenswrapper[4677]: I1203 14:00:02.988552 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-config-volume\") pod \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " Dec 03 14:00:02 crc kubenswrapper[4677]: I1203 14:00:02.988613 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwp7g\" (UniqueName: \"kubernetes.io/projected/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-kube-api-access-rwp7g\") pod \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " Dec 03 14:00:02 crc kubenswrapper[4677]: I1203 14:00:02.988638 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-secret-volume\") pod \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\" (UID: \"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8\") " Dec 03 14:00:02 crc kubenswrapper[4677]: I1203 14:00:02.989557 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-config-volume" (OuterVolumeSpecName: "config-volume") pod "f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" (UID: "f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:00:02 crc kubenswrapper[4677]: I1203 14:00:02.996693 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-kube-api-access-rwp7g" (OuterVolumeSpecName: "kube-api-access-rwp7g") pod "f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" (UID: "f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8"). InnerVolumeSpecName "kube-api-access-rwp7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:00:02 crc kubenswrapper[4677]: I1203 14:00:02.997766 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" (UID: "f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:00:03 crc kubenswrapper[4677]: I1203 14:00:03.090154 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwp7g\" (UniqueName: \"kubernetes.io/projected/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-kube-api-access-rwp7g\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4677]: I1203 14:00:03.090345 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4677]: I1203 14:00:03.091086 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4677]: I1203 14:00:03.600639 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" event={"ID":"f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8","Type":"ContainerDied","Data":"cec68dcf1c4cc6c1f3c74db6e714aeca110f657ca2365dca17cb2695f4460b97"} Dec 03 14:00:03 crc kubenswrapper[4677]: I1203 14:00:03.600715 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cec68dcf1c4cc6c1f3c74db6e714aeca110f657ca2365dca17cb2695f4460b97" Dec 03 14:00:03 crc kubenswrapper[4677]: I1203 14:00:03.600756 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z" Dec 03 14:00:08 crc kubenswrapper[4677]: I1203 14:00:08.437426 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:00:08 crc kubenswrapper[4677]: I1203 14:00:08.437892 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.692837 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pwkvm"] Dec 03 14:00:23 crc kubenswrapper[4677]: E1203 14:00:23.693597 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" containerName="collect-profiles" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.693609 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" containerName="collect-profiles" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.693721 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" containerName="collect-profiles" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.694102 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.696109 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.696393 4677 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wj94q" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.696407 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.714850 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pwkvm"] Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.720429 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-b5kfm"] Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.721232 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.727757 4677 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-9xgms" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.734755 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-b5kfm"] Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.745053 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zbfln"] Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.745766 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.748881 4677 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-csz28" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.762125 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zbfln"] Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.855545 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrbvf\" (UniqueName: \"kubernetes.io/projected/ecdcc3d2-b371-4439-93b8-fe9e6a945ed9-kube-api-access-jrbvf\") pod \"cert-manager-webhook-5655c58dd6-zbfln\" (UID: \"ecdcc3d2-b371-4439-93b8-fe9e6a945ed9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.855904 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw642\" (UniqueName: \"kubernetes.io/projected/84174520-f102-4fcb-ae53-bb5dd2218549-kube-api-access-bw642\") pod \"cert-manager-cainjector-7f985d654d-pwkvm\" (UID: \"84174520-f102-4fcb-ae53-bb5dd2218549\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.855972 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5gk4\" (UniqueName: \"kubernetes.io/projected/ccc13e6c-e114-482c-8882-fed3fd6324a0-kube-api-access-p5gk4\") pod \"cert-manager-5b446d88c5-b5kfm\" (UID: \"ccc13e6c-e114-482c-8882-fed3fd6324a0\") " pod="cert-manager/cert-manager-5b446d88c5-b5kfm" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.957184 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrbvf\" (UniqueName: \"kubernetes.io/projected/ecdcc3d2-b371-4439-93b8-fe9e6a945ed9-kube-api-access-jrbvf\") pod \"cert-manager-webhook-5655c58dd6-zbfln\" (UID: \"ecdcc3d2-b371-4439-93b8-fe9e6a945ed9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.957261 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw642\" (UniqueName: \"kubernetes.io/projected/84174520-f102-4fcb-ae53-bb5dd2218549-kube-api-access-bw642\") pod \"cert-manager-cainjector-7f985d654d-pwkvm\" (UID: \"84174520-f102-4fcb-ae53-bb5dd2218549\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.957294 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5gk4\" (UniqueName: \"kubernetes.io/projected/ccc13e6c-e114-482c-8882-fed3fd6324a0-kube-api-access-p5gk4\") pod \"cert-manager-5b446d88c5-b5kfm\" (UID: \"ccc13e6c-e114-482c-8882-fed3fd6324a0\") " pod="cert-manager/cert-manager-5b446d88c5-b5kfm" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.981470 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrbvf\" (UniqueName: \"kubernetes.io/projected/ecdcc3d2-b371-4439-93b8-fe9e6a945ed9-kube-api-access-jrbvf\") pod \"cert-manager-webhook-5655c58dd6-zbfln\" (UID: \"ecdcc3d2-b371-4439-93b8-fe9e6a945ed9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.981540 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5gk4\" (UniqueName: \"kubernetes.io/projected/ccc13e6c-e114-482c-8882-fed3fd6324a0-kube-api-access-p5gk4\") pod \"cert-manager-5b446d88c5-b5kfm\" (UID: \"ccc13e6c-e114-482c-8882-fed3fd6324a0\") " pod="cert-manager/cert-manager-5b446d88c5-b5kfm" Dec 03 14:00:23 crc kubenswrapper[4677]: I1203 14:00:23.983286 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw642\" (UniqueName: \"kubernetes.io/projected/84174520-f102-4fcb-ae53-bb5dd2218549-kube-api-access-bw642\") pod \"cert-manager-cainjector-7f985d654d-pwkvm\" (UID: \"84174520-f102-4fcb-ae53-bb5dd2218549\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.012205 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.036012 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.060110 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.280620 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pwkvm"] Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.293494 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.537570 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-b5kfm"] Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.548984 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zbfln"] Dec 03 14:00:24 crc kubenswrapper[4677]: W1203 14:00:24.552933 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecdcc3d2_b371_4439_93b8_fe9e6a945ed9.slice/crio-b754f92d3bdb52f7f907f09141cdb1b97c6073ff3b1483238dab5448b91ae9de WatchSource:0}: Error finding container b754f92d3bdb52f7f907f09141cdb1b97c6073ff3b1483238dab5448b91ae9de: Status 404 returned error can't find the container with id b754f92d3bdb52f7f907f09141cdb1b97c6073ff3b1483238dab5448b91ae9de Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.717056 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" event={"ID":"ccc13e6c-e114-482c-8882-fed3fd6324a0","Type":"ContainerStarted","Data":"c5e6a68b493c84dd5d2179604ab7261e7d1af48bba0bd37eeaf21fe4144bfd63"} Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.718197 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" event={"ID":"84174520-f102-4fcb-ae53-bb5dd2218549","Type":"ContainerStarted","Data":"da77faaf76597bb473ef4374a7067c529a100827c1b8e9518bb1c30209dfcdaf"} Dec 03 14:00:24 crc kubenswrapper[4677]: I1203 14:00:24.719159 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" event={"ID":"ecdcc3d2-b371-4439-93b8-fe9e6a945ed9","Type":"ContainerStarted","Data":"b754f92d3bdb52f7f907f09141cdb1b97c6073ff3b1483238dab5448b91ae9de"} Dec 03 14:00:28 crc kubenswrapper[4677]: I1203 14:00:28.744571 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" event={"ID":"84174520-f102-4fcb-ae53-bb5dd2218549","Type":"ContainerStarted","Data":"3434d213ab2f218072c01a8c350e6af3de3f41f66bbb8bd4bd1ea6784401f18d"} Dec 03 14:00:28 crc kubenswrapper[4677]: I1203 14:00:28.765999 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-pwkvm" podStartSLOduration=2.209601275 podStartE2EDuration="5.76597654s" podCreationTimestamp="2025-12-03 14:00:23 +0000 UTC" firstStartedPulling="2025-12-03 14:00:24.293237036 +0000 UTC m=+815.039569491" lastFinishedPulling="2025-12-03 14:00:27.849612301 +0000 UTC m=+818.595944756" observedRunningTime="2025-12-03 14:00:28.761194051 +0000 UTC m=+819.507526516" watchObservedRunningTime="2025-12-03 14:00:28.76597654 +0000 UTC m=+819.512309015" Dec 03 14:00:29 crc kubenswrapper[4677]: I1203 14:00:29.751656 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" event={"ID":"ecdcc3d2-b371-4439-93b8-fe9e6a945ed9","Type":"ContainerStarted","Data":"bd471e53ea2d97425b83ae80299ab192e993b34d3cfd1d35ad872f139d35201a"} Dec 03 14:00:29 crc kubenswrapper[4677]: I1203 14:00:29.751965 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" Dec 03 14:00:29 crc kubenswrapper[4677]: I1203 14:00:29.754208 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" event={"ID":"ccc13e6c-e114-482c-8882-fed3fd6324a0","Type":"ContainerStarted","Data":"95df142bc04fb3953490cbcd156763125441ba606c9065241487edcb0e742939"} Dec 03 14:00:29 crc kubenswrapper[4677]: I1203 14:00:29.770387 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" podStartSLOduration=2.054884889 podStartE2EDuration="6.770369283s" podCreationTimestamp="2025-12-03 14:00:23 +0000 UTC" firstStartedPulling="2025-12-03 14:00:24.553577239 +0000 UTC m=+815.299909694" lastFinishedPulling="2025-12-03 14:00:29.269061633 +0000 UTC m=+820.015394088" observedRunningTime="2025-12-03 14:00:29.768068346 +0000 UTC m=+820.514400831" watchObservedRunningTime="2025-12-03 14:00:29.770369283 +0000 UTC m=+820.516701758" Dec 03 14:00:29 crc kubenswrapper[4677]: I1203 14:00:29.786420 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" podStartSLOduration=2.041523369 podStartE2EDuration="6.786403403s" podCreationTimestamp="2025-12-03 14:00:23 +0000 UTC" firstStartedPulling="2025-12-03 14:00:24.547290854 +0000 UTC m=+815.293623309" lastFinishedPulling="2025-12-03 14:00:29.292170888 +0000 UTC m=+820.038503343" observedRunningTime="2025-12-03 14:00:29.782783636 +0000 UTC m=+820.529116121" watchObservedRunningTime="2025-12-03 14:00:29.786403403 +0000 UTC m=+820.532735858" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.063922 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.072864 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pm7pz"] Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.073560 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-controller" containerID="cri-o://7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.073815 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="sbdb" containerID="cri-o://73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.073942 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="nbdb" containerID="cri-o://15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.074105 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="northd" containerID="cri-o://2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.074211 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.074303 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-node" containerID="cri-o://aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.074395 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-acl-logging" containerID="cri-o://8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.115251 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" containerID="cri-o://aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" gracePeriod=30 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.441679 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/3.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.444057 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovn-acl-logging/0.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.444701 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovn-controller/0.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.445094 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.505795 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tsklb"] Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.506337 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="sbdb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.506427 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="sbdb" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.506501 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.506571 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.506638 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507108 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507149 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-node" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507162 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-node" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507190 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="northd" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507201 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="northd" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507221 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kubecfg-setup" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507230 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kubecfg-setup" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507243 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="nbdb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507251 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="nbdb" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507265 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507274 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507291 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507303 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507318 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-acl-logging" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507327 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-acl-logging" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507339 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507348 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507358 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507366 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507583 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="nbdb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507597 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="northd" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507612 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507623 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-acl-logging" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507636 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-node" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507649 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507657 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovn-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507672 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507681 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="sbdb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507690 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: E1203 14:00:34.507825 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.507836 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.508003 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.508020 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerName="ovnkube-controller" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.510312 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606572 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-config\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606661 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606699 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-kubelet\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606726 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-netd\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606759 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaf2e40d-0316-4380-961f-8039e0674f2c-ovn-node-metrics-cert\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606790 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-var-lib-openvswitch\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606818 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-script-lib\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606829 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606825 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606873 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606825 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606842 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-openvswitch\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606874 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607017 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.606938 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-systemd-units\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607110 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-systemd\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607143 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-netns\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607168 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-ovn\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607186 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-etc-openvswitch\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607211 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-env-overrides\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607229 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607242 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lxth\" (UniqueName: \"kubernetes.io/projected/eaf2e40d-0316-4380-961f-8039e0674f2c-kube-api-access-9lxth\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607269 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607254 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607289 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-ovn-kubernetes\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607340 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607342 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-slash\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607363 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-slash" (OuterVolumeSpecName: "host-slash") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607372 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-node-log\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607388 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-log-socket\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607406 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-bin\") pod \"eaf2e40d-0316-4380-961f-8039e0674f2c\" (UID: \"eaf2e40d-0316-4380-961f-8039e0674f2c\") " Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607535 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-kubelet\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607555 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-etc-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607583 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-ovn\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607616 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-log-socket\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607580 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607647 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-log-socket" (OuterVolumeSpecName: "log-socket") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607570 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607607 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607629 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-node-log" (OuterVolumeSpecName: "node-log") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607677 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607695 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovnkube-script-lib\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607777 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-node-log\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607831 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-run-netns\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607848 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-env-overrides\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607864 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607891 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npcbl\" (UniqueName: \"kubernetes.io/projected/75c85ad3-d3cf-4881-9d33-86481c1e6214-kube-api-access-npcbl\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607927 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-run-ovn-kubernetes\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607969 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-systemd\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.607994 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608051 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-cni-bin\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608115 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovn-node-metrics-cert\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608197 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-cni-netd\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608230 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-systemd-units\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608258 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-slash\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608292 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovnkube-config\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608370 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-var-lib-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608445 4677 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608461 4677 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608474 4677 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608486 4677 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608498 4677 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608509 4677 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608521 4677 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608533 4677 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608548 4677 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608561 4677 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608573 4677 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608584 4677 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608594 4677 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608605 4677 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608616 4677 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaf2e40d-0316-4380-961f-8039e0674f2c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608633 4677 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.608645 4677 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.612501 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf2e40d-0316-4380-961f-8039e0674f2c-kube-api-access-9lxth" (OuterVolumeSpecName: "kube-api-access-9lxth") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "kube-api-access-9lxth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.612759 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf2e40d-0316-4380-961f-8039e0674f2c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.622419 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "eaf2e40d-0316-4380-961f-8039e0674f2c" (UID: "eaf2e40d-0316-4380-961f-8039e0674f2c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709488 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-kubelet\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709540 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-etc-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709565 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-ovn\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709592 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-log-socket\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709611 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709641 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovnkube-script-lib\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709659 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-node-log\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709679 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-env-overrides\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709698 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-run-netns\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709715 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npcbl\" (UniqueName: \"kubernetes.io/projected/75c85ad3-d3cf-4881-9d33-86481c1e6214-kube-api-access-npcbl\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709741 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-run-ovn-kubernetes\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709731 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-etc-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709828 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-systemd\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709759 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-systemd\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709873 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-ovn\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709898 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-log-socket\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709914 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709925 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.709983 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-cni-bin\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710043 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovn-node-metrics-cert\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710121 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-cni-netd\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710157 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-systemd-units\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710199 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-slash\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710240 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovnkube-config\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710338 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-var-lib-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710449 4677 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eaf2e40d-0316-4380-961f-8039e0674f2c-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710475 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lxth\" (UniqueName: \"kubernetes.io/projected/eaf2e40d-0316-4380-961f-8039e0674f2c-kube-api-access-9lxth\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710502 4677 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaf2e40d-0316-4380-961f-8039e0674f2c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710558 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-var-lib-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710614 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-kubelet\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710648 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovnkube-script-lib\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710659 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-run-openvswitch\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710703 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-node-log\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.710749 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-cni-bin\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.711014 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-env-overrides\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.711060 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-run-netns\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.711385 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-run-ovn-kubernetes\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.711419 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-systemd-units\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.711445 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-cni-netd\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.711470 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/75c85ad3-d3cf-4881-9d33-86481c1e6214-host-slash\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.711904 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovnkube-config\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.715071 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75c85ad3-d3cf-4881-9d33-86481c1e6214-ovn-node-metrics-cert\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.729867 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npcbl\" (UniqueName: \"kubernetes.io/projected/75c85ad3-d3cf-4881-9d33-86481c1e6214-kube-api-access-npcbl\") pod \"ovnkube-node-tsklb\" (UID: \"75c85ad3-d3cf-4881-9d33-86481c1e6214\") " pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.785022 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovnkube-controller/3.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.788090 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovn-acl-logging/0.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.788650 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pm7pz_eaf2e40d-0316-4380-961f-8039e0674f2c/ovn-controller/0.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789247 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" exitCode=0 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789272 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" exitCode=0 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789281 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" exitCode=0 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789287 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" exitCode=0 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789294 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" exitCode=0 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789303 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" exitCode=0 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789309 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" exitCode=143 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789317 4677 generic.go:334] "Generic (PLEG): container finished" podID="eaf2e40d-0316-4380-961f-8039e0674f2c" containerID="7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" exitCode=143 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789356 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789396 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789409 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789419 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789428 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789439 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789449 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789460 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789466 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789473 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789478 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789483 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789488 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789493 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789499 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789491 4677 scope.go:117] "RemoveContainer" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789505 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789611 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789622 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789628 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789634 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789639 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789644 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789649 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789654 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789659 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789664 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789672 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789680 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789686 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789692 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789697 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789702 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789707 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789712 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789717 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789722 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789727 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789734 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" event={"ID":"eaf2e40d-0316-4380-961f-8039e0674f2c","Type":"ContainerDied","Data":"5a56617fdb324443c399a5eaf6621a4a2d5db7865adccb41b08455ba73d584c3"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789743 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789750 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789756 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789762 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789768 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789773 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789779 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789784 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789789 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789795 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.789591 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pm7pz" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.791196 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/2.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.791669 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/1.log" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.791696 4677 generic.go:334] "Generic (PLEG): container finished" podID="511a6f65-7dac-4f37-a15e-3a24339f80f4" containerID="0b27c752b379608243e6b1a21d41b90e4bf7e320e5eab9403c29cd5a04921959" exitCode=2 Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.791717 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerDied","Data":"0b27c752b379608243e6b1a21d41b90e4bf7e320e5eab9403c29cd5a04921959"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.791731 4677 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c"} Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.792250 4677 scope.go:117] "RemoveContainer" containerID="0b27c752b379608243e6b1a21d41b90e4bf7e320e5eab9403c29cd5a04921959" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.819394 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.830178 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.869987 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pm7pz"] Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.882385 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pm7pz"] Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.896139 4677 scope.go:117] "RemoveContainer" containerID="73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.936189 4677 scope.go:117] "RemoveContainer" containerID="15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.958663 4677 scope.go:117] "RemoveContainer" containerID="2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.976145 4677 scope.go:117] "RemoveContainer" containerID="e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" Dec 03 14:00:34 crc kubenswrapper[4677]: I1203 14:00:34.992408 4677 scope.go:117] "RemoveContainer" containerID="aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.011158 4677 scope.go:117] "RemoveContainer" containerID="8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.025912 4677 scope.go:117] "RemoveContainer" containerID="7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.060120 4677 scope.go:117] "RemoveContainer" containerID="d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.108255 4677 scope.go:117] "RemoveContainer" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.110650 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": container with ID starting with aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603 not found: ID does not exist" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.110698 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} err="failed to get container status \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": rpc error: code = NotFound desc = could not find container \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": container with ID starting with aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.110725 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.111308 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": container with ID starting with ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e not found: ID does not exist" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.111340 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} err="failed to get container status \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": rpc error: code = NotFound desc = could not find container \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": container with ID starting with ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.111357 4677 scope.go:117] "RemoveContainer" containerID="73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.111925 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": container with ID starting with 73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f not found: ID does not exist" containerID="73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.111963 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} err="failed to get container status \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": rpc error: code = NotFound desc = could not find container \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": container with ID starting with 73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.111982 4677 scope.go:117] "RemoveContainer" containerID="15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.112326 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": container with ID starting with 15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb not found: ID does not exist" containerID="15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.112358 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} err="failed to get container status \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": rpc error: code = NotFound desc = could not find container \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": container with ID starting with 15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.112379 4677 scope.go:117] "RemoveContainer" containerID="2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.112673 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": container with ID starting with 2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04 not found: ID does not exist" containerID="2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.112725 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} err="failed to get container status \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": rpc error: code = NotFound desc = could not find container \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": container with ID starting with 2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.112758 4677 scope.go:117] "RemoveContainer" containerID="e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.113296 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": container with ID starting with e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883 not found: ID does not exist" containerID="e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.113348 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} err="failed to get container status \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": rpc error: code = NotFound desc = could not find container \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": container with ID starting with e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.113366 4677 scope.go:117] "RemoveContainer" containerID="aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.113621 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": container with ID starting with aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e not found: ID does not exist" containerID="aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.113647 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} err="failed to get container status \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": rpc error: code = NotFound desc = could not find container \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": container with ID starting with aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.113764 4677 scope.go:117] "RemoveContainer" containerID="8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.114220 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": container with ID starting with 8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8 not found: ID does not exist" containerID="8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.114244 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} err="failed to get container status \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": rpc error: code = NotFound desc = could not find container \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": container with ID starting with 8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.114259 4677 scope.go:117] "RemoveContainer" containerID="7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.114708 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": container with ID starting with 7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57 not found: ID does not exist" containerID="7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.114746 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} err="failed to get container status \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": rpc error: code = NotFound desc = could not find container \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": container with ID starting with 7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.114763 4677 scope.go:117] "RemoveContainer" containerID="d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde" Dec 03 14:00:35 crc kubenswrapper[4677]: E1203 14:00:35.115178 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": container with ID starting with d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde not found: ID does not exist" containerID="d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.115205 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} err="failed to get container status \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": rpc error: code = NotFound desc = could not find container \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": container with ID starting with d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.115222 4677 scope.go:117] "RemoveContainer" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.115594 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} err="failed to get container status \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": rpc error: code = NotFound desc = could not find container \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": container with ID starting with aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.115640 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.116024 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} err="failed to get container status \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": rpc error: code = NotFound desc = could not find container \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": container with ID starting with ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.116051 4677 scope.go:117] "RemoveContainer" containerID="73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.116438 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} err="failed to get container status \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": rpc error: code = NotFound desc = could not find container \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": container with ID starting with 73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.116465 4677 scope.go:117] "RemoveContainer" containerID="15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.116797 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} err="failed to get container status \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": rpc error: code = NotFound desc = could not find container \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": container with ID starting with 15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.116822 4677 scope.go:117] "RemoveContainer" containerID="2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.117101 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} err="failed to get container status \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": rpc error: code = NotFound desc = could not find container \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": container with ID starting with 2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.117123 4677 scope.go:117] "RemoveContainer" containerID="e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.117437 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} err="failed to get container status \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": rpc error: code = NotFound desc = could not find container \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": container with ID starting with e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.117459 4677 scope.go:117] "RemoveContainer" containerID="aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.117877 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} err="failed to get container status \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": rpc error: code = NotFound desc = could not find container \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": container with ID starting with aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.117917 4677 scope.go:117] "RemoveContainer" containerID="8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.118252 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} err="failed to get container status \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": rpc error: code = NotFound desc = could not find container \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": container with ID starting with 8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.118276 4677 scope.go:117] "RemoveContainer" containerID="7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.118716 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} err="failed to get container status \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": rpc error: code = NotFound desc = could not find container \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": container with ID starting with 7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.118747 4677 scope.go:117] "RemoveContainer" containerID="d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.119927 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} err="failed to get container status \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": rpc error: code = NotFound desc = could not find container \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": container with ID starting with d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.119968 4677 scope.go:117] "RemoveContainer" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.120355 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} err="failed to get container status \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": rpc error: code = NotFound desc = could not find container \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": container with ID starting with aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.120377 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.120889 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} err="failed to get container status \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": rpc error: code = NotFound desc = could not find container \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": container with ID starting with ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.120965 4677 scope.go:117] "RemoveContainer" containerID="73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.121316 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} err="failed to get container status \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": rpc error: code = NotFound desc = could not find container \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": container with ID starting with 73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.121343 4677 scope.go:117] "RemoveContainer" containerID="15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.121685 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} err="failed to get container status \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": rpc error: code = NotFound desc = could not find container \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": container with ID starting with 15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.121715 4677 scope.go:117] "RemoveContainer" containerID="2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.122155 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} err="failed to get container status \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": rpc error: code = NotFound desc = could not find container \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": container with ID starting with 2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.122181 4677 scope.go:117] "RemoveContainer" containerID="e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.122493 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} err="failed to get container status \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": rpc error: code = NotFound desc = could not find container \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": container with ID starting with e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.122522 4677 scope.go:117] "RemoveContainer" containerID="aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.122875 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} err="failed to get container status \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": rpc error: code = NotFound desc = could not find container \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": container with ID starting with aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.122903 4677 scope.go:117] "RemoveContainer" containerID="8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.123421 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} err="failed to get container status \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": rpc error: code = NotFound desc = could not find container \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": container with ID starting with 8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.123460 4677 scope.go:117] "RemoveContainer" containerID="7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.123898 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} err="failed to get container status \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": rpc error: code = NotFound desc = could not find container \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": container with ID starting with 7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.123926 4677 scope.go:117] "RemoveContainer" containerID="d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.124495 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} err="failed to get container status \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": rpc error: code = NotFound desc = could not find container \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": container with ID starting with d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.124526 4677 scope.go:117] "RemoveContainer" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.124974 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} err="failed to get container status \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": rpc error: code = NotFound desc = could not find container \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": container with ID starting with aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.125002 4677 scope.go:117] "RemoveContainer" containerID="ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.125288 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e"} err="failed to get container status \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": rpc error: code = NotFound desc = could not find container \"ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e\": container with ID starting with ab6b4c89546742462cf8d19bbd1984b602d44946756bee48f6c4a5a3c4d7b99e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.125317 4677 scope.go:117] "RemoveContainer" containerID="73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.125720 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f"} err="failed to get container status \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": rpc error: code = NotFound desc = could not find container \"73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f\": container with ID starting with 73aeeac623db58327b4b8919f482dac84308b80078e944e992fea20de814978f not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.125753 4677 scope.go:117] "RemoveContainer" containerID="15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.126177 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb"} err="failed to get container status \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": rpc error: code = NotFound desc = could not find container \"15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb\": container with ID starting with 15ce34c16d9bbab86b06ebe3f84c261331097380f528f3896fc3ea73127265bb not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.126208 4677 scope.go:117] "RemoveContainer" containerID="2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.126610 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04"} err="failed to get container status \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": rpc error: code = NotFound desc = could not find container \"2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04\": container with ID starting with 2ebb4493f2e6d230fc64df0e57e267ca96095f8c9fd94bcc32523c2ee3883e04 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.126636 4677 scope.go:117] "RemoveContainer" containerID="e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.126999 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883"} err="failed to get container status \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": rpc error: code = NotFound desc = could not find container \"e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883\": container with ID starting with e6e9d71a6590d02aed28ab97b25e2bc278edc43e1ec5b3fc2fcd5d348d6db883 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.127026 4677 scope.go:117] "RemoveContainer" containerID="aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.127323 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e"} err="failed to get container status \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": rpc error: code = NotFound desc = could not find container \"aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e\": container with ID starting with aabae4d8ef13394a4be12a0972bad1a8737c4089e8e74206245ce04c88a9d88e not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.127350 4677 scope.go:117] "RemoveContainer" containerID="8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.127735 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8"} err="failed to get container status \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": rpc error: code = NotFound desc = could not find container \"8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8\": container with ID starting with 8c4bfc247efce6a9c61d0b3c752dfdb173cf5c101983600014c950db4e68ade8 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.127758 4677 scope.go:117] "RemoveContainer" containerID="7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.128084 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57"} err="failed to get container status \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": rpc error: code = NotFound desc = could not find container \"7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57\": container with ID starting with 7eb14e0a638f83fd5dc4db322255286e9a9cd475338f5202eb1af7eca9b5fd57 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.128110 4677 scope.go:117] "RemoveContainer" containerID="d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.128442 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde"} err="failed to get container status \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": rpc error: code = NotFound desc = could not find container \"d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde\": container with ID starting with d9ab1540e7473bea9ec1aad80df89c8e5b09690ff3c73e47d8bd7dad3ffa4dde not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.128464 4677 scope.go:117] "RemoveContainer" containerID="aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.128915 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603"} err="failed to get container status \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": rpc error: code = NotFound desc = could not find container \"aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603\": container with ID starting with aa75a2bd761c72bbc2e627b679ed34ae07b94632ace620e44763ab8391817603 not found: ID does not exist" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.798373 4677 generic.go:334] "Generic (PLEG): container finished" podID="75c85ad3-d3cf-4881-9d33-86481c1e6214" containerID="2646709a2fc4c4522be5fc6a660d29064c31c97a69cf1bf208ea12a053d49477" exitCode=0 Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.798459 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerDied","Data":"2646709a2fc4c4522be5fc6a660d29064c31c97a69cf1bf208ea12a053d49477"} Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.798525 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"c4c70a8526058e01fa2e432232ebda903b2c01c7f4c8486c2f5857f888124ad6"} Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.800551 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/2.log" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.801246 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/1.log" Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.801335 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7nch7" event={"ID":"511a6f65-7dac-4f37-a15e-3a24339f80f4","Type":"ContainerStarted","Data":"ba9d126439d00fe63c37db46d48391c70968281ba18d1b70eca2e26c9a8efc38"} Dec 03 14:00:35 crc kubenswrapper[4677]: I1203 14:00:35.983355 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf2e40d-0316-4380-961f-8039e0674f2c" path="/var/lib/kubelet/pods/eaf2e40d-0316-4380-961f-8039e0674f2c/volumes" Dec 03 14:00:36 crc kubenswrapper[4677]: I1203 14:00:36.809991 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"e91ff3ea3f8150651f065ec378d2355dfb348ab61b4920a052f732a797875bc9"} Dec 03 14:00:37 crc kubenswrapper[4677]: I1203 14:00:37.819868 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"d6f82a7b07f291ced65b2a982475195b2d8eacc59709a5325369af074f91f049"} Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.437724 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.437780 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.437816 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.438329 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"66093cdea49042d1d130e4c423008a09d5d1d6840fa98fa247eeb0b6e0fa1441"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.438387 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://66093cdea49042d1d130e4c423008a09d5d1d6840fa98fa247eeb0b6e0fa1441" gracePeriod=600 Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.830525 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"c6c04fdeff8c1406f5ec939f173fbfdbc604c434ac55dfa9150ef32dd3130793"} Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.830900 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"687d552df7bb074185f45b62a4a074d7687da72568883e4bb36fe86773583f5d"} Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.830917 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"b39a262ed37c70b396e3407436ca271e8bf0cd6d1488a10006a53a428fcd9e14"} Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.830928 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"a4b07e303baf6343229498ffefa4631d07c850198867913f0f8c8e0e4938c179"} Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.833123 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="66093cdea49042d1d130e4c423008a09d5d1d6840fa98fa247eeb0b6e0fa1441" exitCode=0 Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.833172 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"66093cdea49042d1d130e4c423008a09d5d1d6840fa98fa247eeb0b6e0fa1441"} Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.833204 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"1186ac74eab5d1036a6f0e4e1fb80c6c8725f92e6867eb61894f1292d51d689c"} Dec 03 14:00:38 crc kubenswrapper[4677]: I1203 14:00:38.833225 4677 scope.go:117] "RemoveContainer" containerID="25cc84bb4570658cc4097ad0606cd34b6dcadcbff9d0daecfbed41321f21143e" Dec 03 14:00:41 crc kubenswrapper[4677]: I1203 14:00:41.853631 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"63ff1ca22569d48b1c3c0bd74e3fa35e94c5182a493bd26723b78b45164904af"} Dec 03 14:00:44 crc kubenswrapper[4677]: I1203 14:00:44.873730 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" event={"ID":"75c85ad3-d3cf-4881-9d33-86481c1e6214","Type":"ContainerStarted","Data":"eebe444b084a02220244260473ac1e3b69c4df1e3b2162513ff18a188018ddfb"} Dec 03 14:00:44 crc kubenswrapper[4677]: I1203 14:00:44.874382 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:44 crc kubenswrapper[4677]: I1203 14:00:44.874405 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:44 crc kubenswrapper[4677]: I1203 14:00:44.903988 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" podStartSLOduration=10.903969528 podStartE2EDuration="10.903969528s" podCreationTimestamp="2025-12-03 14:00:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:00:44.900351603 +0000 UTC m=+835.646684068" watchObservedRunningTime="2025-12-03 14:00:44.903969528 +0000 UTC m=+835.650301983" Dec 03 14:00:44 crc kubenswrapper[4677]: I1203 14:00:44.910884 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:45 crc kubenswrapper[4677]: I1203 14:00:45.879593 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:45 crc kubenswrapper[4677]: I1203 14:00:45.908305 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:00:50 crc kubenswrapper[4677]: I1203 14:00:50.411487 4677 scope.go:117] "RemoveContainer" containerID="6930fedef2ca3fdcd4a81f33e38e8173524d99d9cd1eaf5e2b7d9292d517cb3c" Dec 03 14:00:50 crc kubenswrapper[4677]: I1203 14:00:50.911009 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7nch7_511a6f65-7dac-4f37-a15e-3a24339f80f4/kube-multus/2.log" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.847482 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv"] Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.849491 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.851286 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.859142 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv"] Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.866406 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tsklb" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.887241 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.887299 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8n72\" (UniqueName: \"kubernetes.io/projected/82e2ca22-05b0-4755-a323-1ca6afd42937-kube-api-access-r8n72\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.887339 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.988487 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.988618 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.988653 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8n72\" (UniqueName: \"kubernetes.io/projected/82e2ca22-05b0-4755-a323-1ca6afd42937-kube-api-access-r8n72\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.989191 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:04 crc kubenswrapper[4677]: I1203 14:01:04.989470 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:05 crc kubenswrapper[4677]: I1203 14:01:05.010667 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8n72\" (UniqueName: \"kubernetes.io/projected/82e2ca22-05b0-4755-a323-1ca6afd42937-kube-api-access-r8n72\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:05 crc kubenswrapper[4677]: I1203 14:01:05.168843 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:05 crc kubenswrapper[4677]: I1203 14:01:05.350461 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv"] Dec 03 14:01:05 crc kubenswrapper[4677]: W1203 14:01:05.356210 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82e2ca22_05b0_4755_a323_1ca6afd42937.slice/crio-f7281e82b5c33454dcd7d37397a1bed82e16d71b1bc3a8d643949dc1c68876f8 WatchSource:0}: Error finding container f7281e82b5c33454dcd7d37397a1bed82e16d71b1bc3a8d643949dc1c68876f8: Status 404 returned error can't find the container with id f7281e82b5c33454dcd7d37397a1bed82e16d71b1bc3a8d643949dc1c68876f8 Dec 03 14:01:05 crc kubenswrapper[4677]: I1203 14:01:05.995338 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" event={"ID":"82e2ca22-05b0-4755-a323-1ca6afd42937","Type":"ContainerStarted","Data":"8cf5418ab00c8ec0acbeaf7acf774a3b7fbdfe343fe0b15ebb5ac90a62ab35fa"} Dec 03 14:01:05 crc kubenswrapper[4677]: I1203 14:01:05.995712 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" event={"ID":"82e2ca22-05b0-4755-a323-1ca6afd42937","Type":"ContainerStarted","Data":"f7281e82b5c33454dcd7d37397a1bed82e16d71b1bc3a8d643949dc1c68876f8"} Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.003738 4677 generic.go:334] "Generic (PLEG): container finished" podID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerID="8cf5418ab00c8ec0acbeaf7acf774a3b7fbdfe343fe0b15ebb5ac90a62ab35fa" exitCode=0 Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.003934 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" event={"ID":"82e2ca22-05b0-4755-a323-1ca6afd42937","Type":"ContainerDied","Data":"8cf5418ab00c8ec0acbeaf7acf774a3b7fbdfe343fe0b15ebb5ac90a62ab35fa"} Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.194207 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-785mp"] Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.202303 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-785mp"] Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.202433 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.317555 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwbwf\" (UniqueName: \"kubernetes.io/projected/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-kube-api-access-qwbwf\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.317617 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-catalog-content\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.317733 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-utilities\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.419345 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-utilities\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.419711 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwbwf\" (UniqueName: \"kubernetes.io/projected/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-kube-api-access-qwbwf\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.419749 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-catalog-content\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.419911 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-utilities\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.420101 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-catalog-content\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.448152 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwbwf\" (UniqueName: \"kubernetes.io/projected/68ad29ee-1d1c-41ac-8679-9b30bd41e1a6-kube-api-access-qwbwf\") pod \"redhat-operators-785mp\" (UID: \"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6\") " pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.548421 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:07 crc kubenswrapper[4677]: I1203 14:01:07.763897 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-785mp"] Dec 03 14:01:08 crc kubenswrapper[4677]: I1203 14:01:08.028195 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-785mp" event={"ID":"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6","Type":"ContainerStarted","Data":"44ae557bbb056f83d28379afb001383fd83b029f3afce269f782882acabe9cd6"} Dec 03 14:01:08 crc kubenswrapper[4677]: I1203 14:01:08.028434 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-785mp" event={"ID":"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6","Type":"ContainerStarted","Data":"a78878ca20fb476d7c4e69b2f623162fbcdb44e707c092e431e077c5dc3230e6"} Dec 03 14:01:09 crc kubenswrapper[4677]: I1203 14:01:09.036012 4677 generic.go:334] "Generic (PLEG): container finished" podID="68ad29ee-1d1c-41ac-8679-9b30bd41e1a6" containerID="44ae557bbb056f83d28379afb001383fd83b029f3afce269f782882acabe9cd6" exitCode=0 Dec 03 14:01:09 crc kubenswrapper[4677]: I1203 14:01:09.036068 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-785mp" event={"ID":"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6","Type":"ContainerDied","Data":"44ae557bbb056f83d28379afb001383fd83b029f3afce269f782882acabe9cd6"} Dec 03 14:01:12 crc kubenswrapper[4677]: I1203 14:01:12.055597 4677 generic.go:334] "Generic (PLEG): container finished" podID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerID="94cdc414b6f788093fd33e10b1b721b37226f86d526f87971cea43dce3eb5d2b" exitCode=0 Dec 03 14:01:12 crc kubenswrapper[4677]: I1203 14:01:12.055707 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" event={"ID":"82e2ca22-05b0-4755-a323-1ca6afd42937","Type":"ContainerDied","Data":"94cdc414b6f788093fd33e10b1b721b37226f86d526f87971cea43dce3eb5d2b"} Dec 03 14:01:13 crc kubenswrapper[4677]: I1203 14:01:13.064859 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" event={"ID":"82e2ca22-05b0-4755-a323-1ca6afd42937","Type":"ContainerStarted","Data":"f3e23876e81ac12fcca0603dd313d64bfab8dfbabe5c0f840fa5c0986b3fa463"} Dec 03 14:01:13 crc kubenswrapper[4677]: I1203 14:01:13.085706 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" podStartSLOduration=6.179377223 podStartE2EDuration="9.085687929s" podCreationTimestamp="2025-12-03 14:01:04 +0000 UTC" firstStartedPulling="2025-12-03 14:01:07.007510872 +0000 UTC m=+857.753843327" lastFinishedPulling="2025-12-03 14:01:09.913821558 +0000 UTC m=+860.660154033" observedRunningTime="2025-12-03 14:01:13.083756822 +0000 UTC m=+863.830089297" watchObservedRunningTime="2025-12-03 14:01:13.085687929 +0000 UTC m=+863.832020384" Dec 03 14:01:14 crc kubenswrapper[4677]: I1203 14:01:14.071805 4677 generic.go:334] "Generic (PLEG): container finished" podID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerID="f3e23876e81ac12fcca0603dd313d64bfab8dfbabe5c0f840fa5c0986b3fa463" exitCode=0 Dec 03 14:01:14 crc kubenswrapper[4677]: I1203 14:01:14.071861 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" event={"ID":"82e2ca22-05b0-4755-a323-1ca6afd42937","Type":"ContainerDied","Data":"f3e23876e81ac12fcca0603dd313d64bfab8dfbabe5c0f840fa5c0986b3fa463"} Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.734594 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.890106 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-util\") pod \"82e2ca22-05b0-4755-a323-1ca6afd42937\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.890275 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-bundle\") pod \"82e2ca22-05b0-4755-a323-1ca6afd42937\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.890318 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8n72\" (UniqueName: \"kubernetes.io/projected/82e2ca22-05b0-4755-a323-1ca6afd42937-kube-api-access-r8n72\") pod \"82e2ca22-05b0-4755-a323-1ca6afd42937\" (UID: \"82e2ca22-05b0-4755-a323-1ca6afd42937\") " Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.892789 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-bundle" (OuterVolumeSpecName: "bundle") pod "82e2ca22-05b0-4755-a323-1ca6afd42937" (UID: "82e2ca22-05b0-4755-a323-1ca6afd42937"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.897291 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82e2ca22-05b0-4755-a323-1ca6afd42937-kube-api-access-r8n72" (OuterVolumeSpecName: "kube-api-access-r8n72") pod "82e2ca22-05b0-4755-a323-1ca6afd42937" (UID: "82e2ca22-05b0-4755-a323-1ca6afd42937"). InnerVolumeSpecName "kube-api-access-r8n72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.908423 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-util" (OuterVolumeSpecName: "util") pod "82e2ca22-05b0-4755-a323-1ca6afd42937" (UID: "82e2ca22-05b0-4755-a323-1ca6afd42937"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.991511 4677 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.991554 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8n72\" (UniqueName: \"kubernetes.io/projected/82e2ca22-05b0-4755-a323-1ca6afd42937-kube-api-access-r8n72\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:19 crc kubenswrapper[4677]: I1203 14:01:19.991565 4677 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82e2ca22-05b0-4755-a323-1ca6afd42937-util\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:20 crc kubenswrapper[4677]: I1203 14:01:20.303279 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" event={"ID":"82e2ca22-05b0-4755-a323-1ca6afd42937","Type":"ContainerDied","Data":"f7281e82b5c33454dcd7d37397a1bed82e16d71b1bc3a8d643949dc1c68876f8"} Dec 03 14:01:20 crc kubenswrapper[4677]: I1203 14:01:20.303339 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7281e82b5c33454dcd7d37397a1bed82e16d71b1bc3a8d643949dc1c68876f8" Dec 03 14:01:20 crc kubenswrapper[4677]: I1203 14:01:20.303374 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv" Dec 03 14:01:22 crc kubenswrapper[4677]: I1203 14:01:22.317356 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-785mp" event={"ID":"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6","Type":"ContainerStarted","Data":"a2520a2491749f8d253d4fc15e4ac757a5bab241af38d3dcd0d8d38e781854f9"} Dec 03 14:01:23 crc kubenswrapper[4677]: I1203 14:01:23.324566 4677 generic.go:334] "Generic (PLEG): container finished" podID="68ad29ee-1d1c-41ac-8679-9b30bd41e1a6" containerID="a2520a2491749f8d253d4fc15e4ac757a5bab241af38d3dcd0d8d38e781854f9" exitCode=0 Dec 03 14:01:23 crc kubenswrapper[4677]: I1203 14:01:23.324693 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-785mp" event={"ID":"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6","Type":"ContainerDied","Data":"a2520a2491749f8d253d4fc15e4ac757a5bab241af38d3dcd0d8d38e781854f9"} Dec 03 14:01:24 crc kubenswrapper[4677]: I1203 14:01:24.332201 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-785mp" event={"ID":"68ad29ee-1d1c-41ac-8679-9b30bd41e1a6","Type":"ContainerStarted","Data":"481fce084dcd539e0c0e445945ed43afc56cb4852d1ce621bd921e954e12d617"} Dec 03 14:01:24 crc kubenswrapper[4677]: I1203 14:01:24.353611 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-785mp" podStartSLOduration=3.195529157 podStartE2EDuration="17.353590119s" podCreationTimestamp="2025-12-03 14:01:07 +0000 UTC" firstStartedPulling="2025-12-03 14:01:09.911875061 +0000 UTC m=+860.658207516" lastFinishedPulling="2025-12-03 14:01:24.069936013 +0000 UTC m=+874.816268478" observedRunningTime="2025-12-03 14:01:24.35021783 +0000 UTC m=+875.096550295" watchObservedRunningTime="2025-12-03 14:01:24.353590119 +0000 UTC m=+875.099922574" Dec 03 14:01:27 crc kubenswrapper[4677]: I1203 14:01:27.549349 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:27 crc kubenswrapper[4677]: I1203 14:01:27.549887 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:28 crc kubenswrapper[4677]: I1203 14:01:28.754452 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-785mp" podUID="68ad29ee-1d1c-41ac-8679-9b30bd41e1a6" containerName="registry-server" probeResult="failure" output=< Dec 03 14:01:28 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 14:01:28 crc kubenswrapper[4677]: > Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.071236 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9"] Dec 03 14:01:36 crc kubenswrapper[4677]: E1203 14:01:36.072151 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerName="pull" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.072171 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerName="pull" Dec 03 14:01:36 crc kubenswrapper[4677]: E1203 14:01:36.072197 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerName="util" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.072205 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerName="util" Dec 03 14:01:36 crc kubenswrapper[4677]: E1203 14:01:36.072223 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerName="extract" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.072232 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerName="extract" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.072353 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="82e2ca22-05b0-4755-a323-1ca6afd42937" containerName="extract" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.072826 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.077471 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-g7kj5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.077678 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.077804 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.084705 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.134657 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbt2x\" (UniqueName: \"kubernetes.io/projected/61643dea-c510-4111-8ab1-ec1afc438078-kube-api-access-fbt2x\") pod \"obo-prometheus-operator-668cf9dfbb-pkpj9\" (UID: \"61643dea-c510-4111-8ab1-ec1afc438078\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.196919 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.197763 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.201003 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-kvjf8" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.201259 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.223171 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.226077 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.226987 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.235534 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbt2x\" (UniqueName: \"kubernetes.io/projected/61643dea-c510-4111-8ab1-ec1afc438078-kube-api-access-fbt2x\") pod \"obo-prometheus-operator-668cf9dfbb-pkpj9\" (UID: \"61643dea-c510-4111-8ab1-ec1afc438078\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.235571 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b737ec59-d50e-4179-ac09-1bfa1e51529c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm\" (UID: \"b737ec59-d50e-4179-ac09-1bfa1e51529c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.235636 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b737ec59-d50e-4179-ac09-1bfa1e51529c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm\" (UID: \"b737ec59-d50e-4179-ac09-1bfa1e51529c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.247216 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.312341 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbt2x\" (UniqueName: \"kubernetes.io/projected/61643dea-c510-4111-8ab1-ec1afc438078-kube-api-access-fbt2x\") pod \"obo-prometheus-operator-668cf9dfbb-pkpj9\" (UID: \"61643dea-c510-4111-8ab1-ec1afc438078\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.337734 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16fd0377-c577-4f4e-8e02-5d1e31718895-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g\" (UID: \"16fd0377-c577-4f4e-8e02-5d1e31718895\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.337842 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b737ec59-d50e-4179-ac09-1bfa1e51529c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm\" (UID: \"b737ec59-d50e-4179-ac09-1bfa1e51529c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.337936 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16fd0377-c577-4f4e-8e02-5d1e31718895-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g\" (UID: \"16fd0377-c577-4f4e-8e02-5d1e31718895\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.338016 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b737ec59-d50e-4179-ac09-1bfa1e51529c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm\" (UID: \"b737ec59-d50e-4179-ac09-1bfa1e51529c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.345534 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b737ec59-d50e-4179-ac09-1bfa1e51529c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm\" (UID: \"b737ec59-d50e-4179-ac09-1bfa1e51529c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.353562 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b737ec59-d50e-4179-ac09-1bfa1e51529c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm\" (UID: \"b737ec59-d50e-4179-ac09-1bfa1e51529c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.393196 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-ps6w5"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.393894 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.396232 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.397386 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-28cfr" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.404297 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.417544 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-ps6w5"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.438700 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/65806097-c974-427b-a63c-ed84eaa3b039-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-ps6w5\" (UID: \"65806097-c974-427b-a63c-ed84eaa3b039\") " pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.438794 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16fd0377-c577-4f4e-8e02-5d1e31718895-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g\" (UID: \"16fd0377-c577-4f4e-8e02-5d1e31718895\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.438828 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzrw5\" (UniqueName: \"kubernetes.io/projected/65806097-c974-427b-a63c-ed84eaa3b039-kube-api-access-lzrw5\") pod \"observability-operator-d8bb48f5d-ps6w5\" (UID: \"65806097-c974-427b-a63c-ed84eaa3b039\") " pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.438874 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16fd0377-c577-4f4e-8e02-5d1e31718895-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g\" (UID: \"16fd0377-c577-4f4e-8e02-5d1e31718895\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.453892 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16fd0377-c577-4f4e-8e02-5d1e31718895-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g\" (UID: \"16fd0377-c577-4f4e-8e02-5d1e31718895\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.453998 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16fd0377-c577-4f4e-8e02-5d1e31718895-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g\" (UID: \"16fd0377-c577-4f4e-8e02-5d1e31718895\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.514474 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.539856 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/65806097-c974-427b-a63c-ed84eaa3b039-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-ps6w5\" (UID: \"65806097-c974-427b-a63c-ed84eaa3b039\") " pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.539938 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzrw5\" (UniqueName: \"kubernetes.io/projected/65806097-c974-427b-a63c-ed84eaa3b039-kube-api-access-lzrw5\") pod \"observability-operator-d8bb48f5d-ps6w5\" (UID: \"65806097-c974-427b-a63c-ed84eaa3b039\") " pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.545640 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.545720 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/65806097-c974-427b-a63c-ed84eaa3b039-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-ps6w5\" (UID: \"65806097-c974-427b-a63c-ed84eaa3b039\") " pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.561675 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzrw5\" (UniqueName: \"kubernetes.io/projected/65806097-c974-427b-a63c-ed84eaa3b039-kube-api-access-lzrw5\") pod \"observability-operator-d8bb48f5d-ps6w5\" (UID: \"65806097-c974-427b-a63c-ed84eaa3b039\") " pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.645105 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-lm89l"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.653724 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.690671 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-dtw7c" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.710055 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-lm89l"] Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.726150 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.745003 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl7b2\" (UniqueName: \"kubernetes.io/projected/cfed8170-869b-48c2-8b7f-e69466674721-kube-api-access-zl7b2\") pod \"perses-operator-5446b9c989-lm89l\" (UID: \"cfed8170-869b-48c2-8b7f-e69466674721\") " pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.745067 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/cfed8170-869b-48c2-8b7f-e69466674721-openshift-service-ca\") pod \"perses-operator-5446b9c989-lm89l\" (UID: \"cfed8170-869b-48c2-8b7f-e69466674721\") " pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.854138 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/cfed8170-869b-48c2-8b7f-e69466674721-openshift-service-ca\") pod \"perses-operator-5446b9c989-lm89l\" (UID: \"cfed8170-869b-48c2-8b7f-e69466674721\") " pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.854258 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl7b2\" (UniqueName: \"kubernetes.io/projected/cfed8170-869b-48c2-8b7f-e69466674721-kube-api-access-zl7b2\") pod \"perses-operator-5446b9c989-lm89l\" (UID: \"cfed8170-869b-48c2-8b7f-e69466674721\") " pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.855026 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/cfed8170-869b-48c2-8b7f-e69466674721-openshift-service-ca\") pod \"perses-operator-5446b9c989-lm89l\" (UID: \"cfed8170-869b-48c2-8b7f-e69466674721\") " pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.881142 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl7b2\" (UniqueName: \"kubernetes.io/projected/cfed8170-869b-48c2-8b7f-e69466674721-kube-api-access-zl7b2\") pod \"perses-operator-5446b9c989-lm89l\" (UID: \"cfed8170-869b-48c2-8b7f-e69466674721\") " pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:36 crc kubenswrapper[4677]: I1203 14:01:36.935856 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9"] Dec 03 14:01:36 crc kubenswrapper[4677]: W1203 14:01:36.958308 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61643dea_c510_4111_8ab1_ec1afc438078.slice/crio-adcfb3c7c3b07a716153b0a85542c7a3fde37db035b28683b8bf104deb5bb1c5 WatchSource:0}: Error finding container adcfb3c7c3b07a716153b0a85542c7a3fde37db035b28683b8bf104deb5bb1c5: Status 404 returned error can't find the container with id adcfb3c7c3b07a716153b0a85542c7a3fde37db035b28683b8bf104deb5bb1c5 Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.074356 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.218340 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-ps6w5"] Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.364838 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm"] Dec 03 14:01:37 crc kubenswrapper[4677]: W1203 14:01:37.376588 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb737ec59_d50e_4179_ac09_1bfa1e51529c.slice/crio-3f5f70bb8a9e78cfe138c46a4e75f8a88e8dafc812de718b47dba687aed49b8f WatchSource:0}: Error finding container 3f5f70bb8a9e78cfe138c46a4e75f8a88e8dafc812de718b47dba687aed49b8f: Status 404 returned error can't find the container with id 3f5f70bb8a9e78cfe138c46a4e75f8a88e8dafc812de718b47dba687aed49b8f Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.383825 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g"] Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.500519 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-lm89l"] Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.515746 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" event={"ID":"61643dea-c510-4111-8ab1-ec1afc438078","Type":"ContainerStarted","Data":"adcfb3c7c3b07a716153b0a85542c7a3fde37db035b28683b8bf104deb5bb1c5"} Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.517765 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" event={"ID":"16fd0377-c577-4f4e-8e02-5d1e31718895","Type":"ContainerStarted","Data":"9ad6c89e228b41e3130cca4bdaed093fd750bfd143ddc32bcbb1f5d3fd149d86"} Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.518974 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" event={"ID":"65806097-c974-427b-a63c-ed84eaa3b039","Type":"ContainerStarted","Data":"21556b658f9b245cc3458ec95571b3a211f24c4d1ad35983a2ccd00da916fa4b"} Dec 03 14:01:37 crc kubenswrapper[4677]: W1203 14:01:37.519876 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfed8170_869b_48c2_8b7f_e69466674721.slice/crio-adc42138f4e9506f6065244c874500f381f9394850e936b84931ae1ead361687 WatchSource:0}: Error finding container adc42138f4e9506f6065244c874500f381f9394850e936b84931ae1ead361687: Status 404 returned error can't find the container with id adc42138f4e9506f6065244c874500f381f9394850e936b84931ae1ead361687 Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.520128 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" event={"ID":"b737ec59-d50e-4179-ac09-1bfa1e51529c","Type":"ContainerStarted","Data":"3f5f70bb8a9e78cfe138c46a4e75f8a88e8dafc812de718b47dba687aed49b8f"} Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.604509 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:37 crc kubenswrapper[4677]: I1203 14:01:37.687355 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-785mp" Dec 03 14:01:38 crc kubenswrapper[4677]: I1203 14:01:38.222000 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-785mp"] Dec 03 14:01:38 crc kubenswrapper[4677]: I1203 14:01:38.388965 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7n5sl"] Dec 03 14:01:38 crc kubenswrapper[4677]: I1203 14:01:38.389458 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7n5sl" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="registry-server" containerID="cri-o://b1f8053ecee7e87b26fbb81dea6dc971e944c67374d15131201fb65e2ccf77b6" gracePeriod=2 Dec 03 14:01:38 crc kubenswrapper[4677]: I1203 14:01:38.526403 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-lm89l" event={"ID":"cfed8170-869b-48c2-8b7f-e69466674721","Type":"ContainerStarted","Data":"adc42138f4e9506f6065244c874500f381f9394850e936b84931ae1ead361687"} Dec 03 14:01:42 crc kubenswrapper[4677]: I1203 14:01:42.599841 4677 generic.go:334] "Generic (PLEG): container finished" podID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerID="b1f8053ecee7e87b26fbb81dea6dc971e944c67374d15131201fb65e2ccf77b6" exitCode=0 Dec 03 14:01:42 crc kubenswrapper[4677]: I1203 14:01:42.599914 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7n5sl" event={"ID":"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c","Type":"ContainerDied","Data":"b1f8053ecee7e87b26fbb81dea6dc971e944c67374d15131201fb65e2ccf77b6"} Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.033751 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.217244 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-catalog-content\") pod \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.217598 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-utilities\") pod \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.217666 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxq7t\" (UniqueName: \"kubernetes.io/projected/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-kube-api-access-dxq7t\") pod \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\" (UID: \"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c\") " Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.219499 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-utilities" (OuterVolumeSpecName: "utilities") pod "0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" (UID: "0b643fc0-7b4f-4e93-a9eb-32b7fa40353c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.226991 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-kube-api-access-dxq7t" (OuterVolumeSpecName: "kube-api-access-dxq7t") pod "0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" (UID: "0b643fc0-7b4f-4e93-a9eb-32b7fa40353c"). InnerVolumeSpecName "kube-api-access-dxq7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.318645 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxq7t\" (UniqueName: \"kubernetes.io/projected/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-kube-api-access-dxq7t\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.318676 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.339893 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" (UID: "0b643fc0-7b4f-4e93-a9eb-32b7fa40353c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.419363 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.657804 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7n5sl" event={"ID":"0b643fc0-7b4f-4e93-a9eb-32b7fa40353c","Type":"ContainerDied","Data":"2868e966fae95a263f7e580a4f35a4e3e5cd080b0390d80c9124db1ea39bd1fc"} Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.657868 4677 scope.go:117] "RemoveContainer" containerID="b1f8053ecee7e87b26fbb81dea6dc971e944c67374d15131201fb65e2ccf77b6" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.658034 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7n5sl" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.689478 4677 scope.go:117] "RemoveContainer" containerID="0c27748c1cb7898c4f024a01a0138c7f8d63ed2266359d8dc18ac3bfd0e574de" Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.705106 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7n5sl"] Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.718099 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7n5sl"] Dec 03 14:01:44 crc kubenswrapper[4677]: I1203 14:01:44.752847 4677 scope.go:117] "RemoveContainer" containerID="954724ac90d374ece8ed221777465e7e8782cbae90e4ae910b2838e0bcfd12bc" Dec 03 14:01:45 crc kubenswrapper[4677]: I1203 14:01:45.986606 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" path="/var/lib/kubelet/pods/0b643fc0-7b4f-4e93-a9eb-32b7fa40353c/volumes" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.210704 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.211458 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:e718854a7d6ca8accf0fa72db0eb902e46c44d747ad51dc3f06bba0cefaa3c01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:17ea20be390a94ab39f5cdd7f0cbc2498046eebcf77fe3dec9aa288d5c2cf46b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:d972f4faa5e9c121402d23ed85002f26af48ec36b1b71a7489d677b3913d08b4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:91531137fc1dcd740e277e0f65e120a0176a16f788c14c27925b61aa0b792ade,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:a69da8bbca8a28dd2925f864d51cc31cf761b10532c553095ba40b242ef701cb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:897e1bfad1187062725b54d87107bd0155972257a50d8335dd29e1999b828a4f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:95fe5b5746ca8c07ac9217ce2d8ac8e6afad17af210f9d8e0074df1310b209a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:e9d9a89e4d8126a62b1852055482258ee528cac6398dd5d43ebad75ace0f33c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:ec684a0645ceb917b019af7ddba68c3533416e356ab0d0320a30e75ca7ebb31b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:3b9693fcde9b3a9494fb04735b1f7cfd0426f10be820fdc3f024175c0d3df1c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:580606f194180accc8abba099e17a26dca7522ec6d233fa2fdd40312771703e3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:e03777be39e71701935059cd877603874a13ac94daa73219d4e5e545599d78a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:aa47256193cfd2877853878e1ae97d2ab8b8e5deae62b387cbfad02b284d379c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:c595ff56b2cb85514bf4784db6ddb82e4e657e3e708a7fb695fc4997379a94d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:45a4ec2a519bcec99e886aa91596d5356a2414a2bd103baaef9fa7838c672eb2,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lzrw5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-d8bb48f5d-ps6w5_openshift-operators(65806097-c974-427b-a63c-ed84eaa3b039): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.212717 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" podUID="65806097-c974-427b-a63c-ed84eaa3b039" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.719964 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.720202 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zl7b2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-lm89l_openshift-operators(cfed8170-869b-48c2-8b7f-e69466674721): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.721412 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-lm89l" podUID="cfed8170-869b-48c2-8b7f-e69466674721" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.813851 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb\\\"\"" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" podUID="65806097-c974-427b-a63c-ed84eaa3b039" Dec 03 14:02:00 crc kubenswrapper[4677]: E1203 14:02:00.814269 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-lm89l" podUID="cfed8170-869b-48c2-8b7f-e69466674721" Dec 03 14:02:01 crc kubenswrapper[4677]: I1203 14:02:01.818316 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" event={"ID":"16fd0377-c577-4f4e-8e02-5d1e31718895","Type":"ContainerStarted","Data":"9eb5058a412ea54d9a0b4aef22b96ee185bbc7b85fc1d5d2cc098c47cee67f4a"} Dec 03 14:02:01 crc kubenswrapper[4677]: I1203 14:02:01.820995 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" event={"ID":"b737ec59-d50e-4179-ac09-1bfa1e51529c","Type":"ContainerStarted","Data":"608bfc353658765c47882a87ae3aec7ffd2c93747f9c2208bdee454762075a47"} Dec 03 14:02:01 crc kubenswrapper[4677]: I1203 14:02:01.823181 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" event={"ID":"61643dea-c510-4111-8ab1-ec1afc438078","Type":"ContainerStarted","Data":"d2df71c843c59e8a4dc7e42b980f2303829113b15e0b54d0f6f092b89e2bd8fe"} Dec 03 14:02:01 crc kubenswrapper[4677]: I1203 14:02:01.843981 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g" podStartSLOduration=2.436834556 podStartE2EDuration="25.843942467s" podCreationTimestamp="2025-12-03 14:01:36 +0000 UTC" firstStartedPulling="2025-12-03 14:01:37.404309527 +0000 UTC m=+888.150641982" lastFinishedPulling="2025-12-03 14:02:00.811417438 +0000 UTC m=+911.557749893" observedRunningTime="2025-12-03 14:02:01.839433305 +0000 UTC m=+912.585765760" watchObservedRunningTime="2025-12-03 14:02:01.843942467 +0000 UTC m=+912.590274912" Dec 03 14:02:01 crc kubenswrapper[4677]: I1203 14:02:01.888226 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-pkpj9" podStartSLOduration=2.056028035 podStartE2EDuration="25.888204908s" podCreationTimestamp="2025-12-03 14:01:36 +0000 UTC" firstStartedPulling="2025-12-03 14:01:36.968239104 +0000 UTC m=+887.714571559" lastFinishedPulling="2025-12-03 14:02:00.800415957 +0000 UTC m=+911.546748432" observedRunningTime="2025-12-03 14:02:01.886088176 +0000 UTC m=+912.632420651" watchObservedRunningTime="2025-12-03 14:02:01.888204908 +0000 UTC m=+912.634537363" Dec 03 14:02:01 crc kubenswrapper[4677]: I1203 14:02:01.912228 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm" podStartSLOduration=2.5302199290000003 podStartE2EDuration="25.912208807s" podCreationTimestamp="2025-12-03 14:01:36 +0000 UTC" firstStartedPulling="2025-12-03 14:01:37.392322217 +0000 UTC m=+888.138654672" lastFinishedPulling="2025-12-03 14:02:00.774311085 +0000 UTC m=+911.520643550" observedRunningTime="2025-12-03 14:02:01.910195769 +0000 UTC m=+912.656528234" watchObservedRunningTime="2025-12-03 14:02:01.912208807 +0000 UTC m=+912.658541252" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.852547 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-88h22"] Dec 03 14:02:17 crc kubenswrapper[4677]: E1203 14:02:17.853628 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="registry-server" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.853644 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="registry-server" Dec 03 14:02:17 crc kubenswrapper[4677]: E1203 14:02:17.853660 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="extract-utilities" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.853667 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="extract-utilities" Dec 03 14:02:17 crc kubenswrapper[4677]: E1203 14:02:17.853689 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="extract-content" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.853698 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="extract-content" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.853822 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b643fc0-7b4f-4e93-a9eb-32b7fa40353c" containerName="registry-server" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.854844 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.866416 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-88h22"] Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.923781 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" event={"ID":"65806097-c974-427b-a63c-ed84eaa3b039","Type":"ContainerStarted","Data":"7551891c167371c008451ee5a774788c9d6439acda183ed25bc0bb4300696065"} Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.924092 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.925187 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-lm89l" event={"ID":"cfed8170-869b-48c2-8b7f-e69466674721","Type":"ContainerStarted","Data":"2bd77c0e3ff0398cbb1a45140b6615424c6b89d52fcc0d26752ba844e310de81"} Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.925369 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.925777 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.952122 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-catalog-content\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.952202 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-utilities\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.952306 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrqfq\" (UniqueName: \"kubernetes.io/projected/36869ae6-cec0-45e4-ae7e-9a703c52ad77-kube-api-access-jrqfq\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:17 crc kubenswrapper[4677]: I1203 14:02:17.956231 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" podStartSLOduration=2.494425382 podStartE2EDuration="41.956210342s" podCreationTimestamp="2025-12-03 14:01:36 +0000 UTC" firstStartedPulling="2025-12-03 14:01:37.236905341 +0000 UTC m=+887.983237796" lastFinishedPulling="2025-12-03 14:02:16.698690301 +0000 UTC m=+927.445022756" observedRunningTime="2025-12-03 14:02:17.953111361 +0000 UTC m=+928.699443826" watchObservedRunningTime="2025-12-03 14:02:17.956210342 +0000 UTC m=+928.702542807" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.005280 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-lm89l" podStartSLOduration=2.3264321040000002 podStartE2EDuration="42.005253103s" podCreationTimestamp="2025-12-03 14:01:36 +0000 UTC" firstStartedPulling="2025-12-03 14:01:37.522612907 +0000 UTC m=+888.268945362" lastFinishedPulling="2025-12-03 14:02:17.201433906 +0000 UTC m=+927.947766361" observedRunningTime="2025-12-03 14:02:18.00107525 +0000 UTC m=+928.747407725" watchObservedRunningTime="2025-12-03 14:02:18.005253103 +0000 UTC m=+928.751585598" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.053271 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrqfq\" (UniqueName: \"kubernetes.io/projected/36869ae6-cec0-45e4-ae7e-9a703c52ad77-kube-api-access-jrqfq\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.053346 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-catalog-content\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.053368 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-utilities\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.053838 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-utilities\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.053920 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-catalog-content\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.086196 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrqfq\" (UniqueName: \"kubernetes.io/projected/36869ae6-cec0-45e4-ae7e-9a703c52ad77-kube-api-access-jrqfq\") pod \"certified-operators-88h22\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.173209 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.595741 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-88h22"] Dec 03 14:02:18 crc kubenswrapper[4677]: I1203 14:02:18.931178 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88h22" event={"ID":"36869ae6-cec0-45e4-ae7e-9a703c52ad77","Type":"ContainerStarted","Data":"62570f9c8b1fdc31e4f81ea61cebe91993ae5c44bb20225f0402193dc1cbf392"} Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.696931 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gcpbf"] Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.699607 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.709060 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcpbf"] Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.835657 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtb65\" (UniqueName: \"kubernetes.io/projected/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-kube-api-access-dtb65\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.836056 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-utilities\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.836307 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-catalog-content\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.937966 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-catalog-content\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.938035 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtb65\" (UniqueName: \"kubernetes.io/projected/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-kube-api-access-dtb65\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.938092 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-utilities\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.938717 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-utilities\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.938937 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-catalog-content\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.958677 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtb65\" (UniqueName: \"kubernetes.io/projected/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-kube-api-access-dtb65\") pod \"redhat-marketplace-gcpbf\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.963114 4677 generic.go:334] "Generic (PLEG): container finished" podID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerID="508583c9d979c4fe32fa010f8df19cfd51a9077a7da7472d4a4e65b4f1a52698" exitCode=0 Dec 03 14:02:23 crc kubenswrapper[4677]: I1203 14:02:23.963166 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88h22" event={"ID":"36869ae6-cec0-45e4-ae7e-9a703c52ad77","Type":"ContainerDied","Data":"508583c9d979c4fe32fa010f8df19cfd51a9077a7da7472d4a4e65b4f1a52698"} Dec 03 14:02:24 crc kubenswrapper[4677]: I1203 14:02:24.017396 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:24 crc kubenswrapper[4677]: I1203 14:02:24.346928 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcpbf"] Dec 03 14:02:24 crc kubenswrapper[4677]: I1203 14:02:24.972790 4677 generic.go:334] "Generic (PLEG): container finished" podID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerID="b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175" exitCode=0 Dec 03 14:02:24 crc kubenswrapper[4677]: I1203 14:02:24.972867 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcpbf" event={"ID":"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f","Type":"ContainerDied","Data":"b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175"} Dec 03 14:02:24 crc kubenswrapper[4677]: I1203 14:02:24.973422 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcpbf" event={"ID":"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f","Type":"ContainerStarted","Data":"27cac383e36e8da1d31191f6d9b26e2d6b35ba2f725b5b91bbdf9d8c612276a4"} Dec 03 14:02:25 crc kubenswrapper[4677]: I1203 14:02:25.990433 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcpbf" event={"ID":"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f","Type":"ContainerStarted","Data":"bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17"} Dec 03 14:02:25 crc kubenswrapper[4677]: I1203 14:02:25.993653 4677 generic.go:334] "Generic (PLEG): container finished" podID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerID="87bebbc70b3f6a9a0d8a8cf56c9cadf2991b4ea9b9c4d1d384ee1f9c3fa2b5ee" exitCode=0 Dec 03 14:02:25 crc kubenswrapper[4677]: I1203 14:02:25.993684 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88h22" event={"ID":"36869ae6-cec0-45e4-ae7e-9a703c52ad77","Type":"ContainerDied","Data":"87bebbc70b3f6a9a0d8a8cf56c9cadf2991b4ea9b9c4d1d384ee1f9c3fa2b5ee"} Dec 03 14:02:27 crc kubenswrapper[4677]: I1203 14:02:27.000812 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88h22" event={"ID":"36869ae6-cec0-45e4-ae7e-9a703c52ad77","Type":"ContainerStarted","Data":"7ac6eabe63493e1dfb871c2f4556d763841e24e8f1895670ba5145046989c5a0"} Dec 03 14:02:27 crc kubenswrapper[4677]: I1203 14:02:27.002666 4677 generic.go:334] "Generic (PLEG): container finished" podID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerID="bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17" exitCode=0 Dec 03 14:02:27 crc kubenswrapper[4677]: I1203 14:02:27.002696 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcpbf" event={"ID":"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f","Type":"ContainerDied","Data":"bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17"} Dec 03 14:02:27 crc kubenswrapper[4677]: I1203 14:02:27.051566 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-88h22" podStartSLOduration=7.476593499 podStartE2EDuration="10.051543399s" podCreationTimestamp="2025-12-03 14:02:17 +0000 UTC" firstStartedPulling="2025-12-03 14:02:23.964599564 +0000 UTC m=+934.710932019" lastFinishedPulling="2025-12-03 14:02:26.539549464 +0000 UTC m=+937.285881919" observedRunningTime="2025-12-03 14:02:27.027941481 +0000 UTC m=+937.774273956" watchObservedRunningTime="2025-12-03 14:02:27.051543399 +0000 UTC m=+937.797875854" Dec 03 14:02:27 crc kubenswrapper[4677]: I1203 14:02:27.077821 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-lm89l" Dec 03 14:02:28 crc kubenswrapper[4677]: I1203 14:02:28.173723 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:28 crc kubenswrapper[4677]: I1203 14:02:28.173944 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:28 crc kubenswrapper[4677]: I1203 14:02:28.218807 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:29 crc kubenswrapper[4677]: I1203 14:02:29.055409 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcpbf" event={"ID":"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f","Type":"ContainerStarted","Data":"de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083"} Dec 03 14:02:29 crc kubenswrapper[4677]: I1203 14:02:29.077633 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gcpbf" podStartSLOduration=3.298219054 podStartE2EDuration="6.077615648s" podCreationTimestamp="2025-12-03 14:02:23 +0000 UTC" firstStartedPulling="2025-12-03 14:02:24.974726809 +0000 UTC m=+935.721059264" lastFinishedPulling="2025-12-03 14:02:27.754123403 +0000 UTC m=+938.500455858" observedRunningTime="2025-12-03 14:02:29.072702425 +0000 UTC m=+939.819034900" watchObservedRunningTime="2025-12-03 14:02:29.077615648 +0000 UTC m=+939.823948103" Dec 03 14:02:34 crc kubenswrapper[4677]: I1203 14:02:34.018241 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:34 crc kubenswrapper[4677]: I1203 14:02:34.018788 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:34 crc kubenswrapper[4677]: I1203 14:02:34.054976 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:34 crc kubenswrapper[4677]: I1203 14:02:34.113931 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:34 crc kubenswrapper[4677]: I1203 14:02:34.292057 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcpbf"] Dec 03 14:02:36 crc kubenswrapper[4677]: I1203 14:02:36.089749 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gcpbf" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="registry-server" containerID="cri-o://de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083" gracePeriod=2 Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.693187 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.775632 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-utilities\") pod \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.776737 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-utilities" (OuterVolumeSpecName: "utilities") pod "1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" (UID: "1fd86cae-8c5e-4685-b38c-6ce56ed49d1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.775938 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtb65\" (UniqueName: \"kubernetes.io/projected/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-kube-api-access-dtb65\") pod \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.777007 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-catalog-content\") pod \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\" (UID: \"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f\") " Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.777299 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.783668 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-kube-api-access-dtb65" (OuterVolumeSpecName: "kube-api-access-dtb65") pod "1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" (UID: "1fd86cae-8c5e-4685-b38c-6ce56ed49d1f"). InnerVolumeSpecName "kube-api-access-dtb65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.814338 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" (UID: "1fd86cae-8c5e-4685-b38c-6ce56ed49d1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.877938 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtb65\" (UniqueName: \"kubernetes.io/projected/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-kube-api-access-dtb65\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:37 crc kubenswrapper[4677]: I1203 14:02:37.877984 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.106227 4677 generic.go:334] "Generic (PLEG): container finished" podID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerID="de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083" exitCode=0 Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.106291 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcpbf" event={"ID":"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f","Type":"ContainerDied","Data":"de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083"} Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.106374 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcpbf" event={"ID":"1fd86cae-8c5e-4685-b38c-6ce56ed49d1f","Type":"ContainerDied","Data":"27cac383e36e8da1d31191f6d9b26e2d6b35ba2f725b5b91bbdf9d8c612276a4"} Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.106407 4677 scope.go:117] "RemoveContainer" containerID="de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.106541 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcpbf" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.137623 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcpbf"] Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.141338 4677 scope.go:117] "RemoveContainer" containerID="bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.143440 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcpbf"] Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.162327 4677 scope.go:117] "RemoveContainer" containerID="b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.185625 4677 scope.go:117] "RemoveContainer" containerID="de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083" Dec 03 14:02:38 crc kubenswrapper[4677]: E1203 14:02:38.186605 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083\": container with ID starting with de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083 not found: ID does not exist" containerID="de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.186642 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083"} err="failed to get container status \"de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083\": rpc error: code = NotFound desc = could not find container \"de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083\": container with ID starting with de53ed3cd890f43a59cb2914048a90799b5aac608b3cddeca1ad5631410ee083 not found: ID does not exist" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.186668 4677 scope.go:117] "RemoveContainer" containerID="bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17" Dec 03 14:02:38 crc kubenswrapper[4677]: E1203 14:02:38.187114 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17\": container with ID starting with bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17 not found: ID does not exist" containerID="bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.187143 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17"} err="failed to get container status \"bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17\": rpc error: code = NotFound desc = could not find container \"bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17\": container with ID starting with bce1b700b7f7d003cea8f01d07328f605e76c50e7a87cf9e7d674c8e471f8c17 not found: ID does not exist" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.187161 4677 scope.go:117] "RemoveContainer" containerID="b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175" Dec 03 14:02:38 crc kubenswrapper[4677]: E1203 14:02:38.187401 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175\": container with ID starting with b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175 not found: ID does not exist" containerID="b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.187430 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175"} err="failed to get container status \"b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175\": rpc error: code = NotFound desc = could not find container \"b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175\": container with ID starting with b3e56595935ab89b8a22c9911f50e3543a848fd6a8a5f4dcba341c2867d04175 not found: ID does not exist" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.223752 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.437027 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:02:38 crc kubenswrapper[4677]: I1203 14:02:38.437087 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:02:39 crc kubenswrapper[4677]: I1203 14:02:39.922083 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-88h22"] Dec 03 14:02:39 crc kubenswrapper[4677]: I1203 14:02:39.922574 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-88h22" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="registry-server" containerID="cri-o://7ac6eabe63493e1dfb871c2f4556d763841e24e8f1895670ba5145046989c5a0" gracePeriod=2 Dec 03 14:02:39 crc kubenswrapper[4677]: I1203 14:02:39.984725 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" path="/var/lib/kubelet/pods/1fd86cae-8c5e-4685-b38c-6ce56ed49d1f/volumes" Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.124352 4677 generic.go:334] "Generic (PLEG): container finished" podID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerID="7ac6eabe63493e1dfb871c2f4556d763841e24e8f1895670ba5145046989c5a0" exitCode=0 Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.124567 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88h22" event={"ID":"36869ae6-cec0-45e4-ae7e-9a703c52ad77","Type":"ContainerDied","Data":"7ac6eabe63493e1dfb871c2f4556d763841e24e8f1895670ba5145046989c5a0"} Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.200992 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.316666 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-utilities\") pod \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.316796 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrqfq\" (UniqueName: \"kubernetes.io/projected/36869ae6-cec0-45e4-ae7e-9a703c52ad77-kube-api-access-jrqfq\") pod \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.316999 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-catalog-content\") pod \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\" (UID: \"36869ae6-cec0-45e4-ae7e-9a703c52ad77\") " Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.318085 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-utilities" (OuterVolumeSpecName: "utilities") pod "36869ae6-cec0-45e4-ae7e-9a703c52ad77" (UID: "36869ae6-cec0-45e4-ae7e-9a703c52ad77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.326935 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36869ae6-cec0-45e4-ae7e-9a703c52ad77-kube-api-access-jrqfq" (OuterVolumeSpecName: "kube-api-access-jrqfq") pod "36869ae6-cec0-45e4-ae7e-9a703c52ad77" (UID: "36869ae6-cec0-45e4-ae7e-9a703c52ad77"). InnerVolumeSpecName "kube-api-access-jrqfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.368464 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36869ae6-cec0-45e4-ae7e-9a703c52ad77" (UID: "36869ae6-cec0-45e4-ae7e-9a703c52ad77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.418140 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.418176 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36869ae6-cec0-45e4-ae7e-9a703c52ad77-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:41 crc kubenswrapper[4677]: I1203 14:02:41.418186 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrqfq\" (UniqueName: \"kubernetes.io/projected/36869ae6-cec0-45e4-ae7e-9a703c52ad77-kube-api-access-jrqfq\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:42 crc kubenswrapper[4677]: I1203 14:02:42.132273 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88h22" event={"ID":"36869ae6-cec0-45e4-ae7e-9a703c52ad77","Type":"ContainerDied","Data":"62570f9c8b1fdc31e4f81ea61cebe91993ae5c44bb20225f0402193dc1cbf392"} Dec 03 14:02:42 crc kubenswrapper[4677]: I1203 14:02:42.132510 4677 scope.go:117] "RemoveContainer" containerID="7ac6eabe63493e1dfb871c2f4556d763841e24e8f1895670ba5145046989c5a0" Dec 03 14:02:42 crc kubenswrapper[4677]: I1203 14:02:42.132535 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88h22" Dec 03 14:02:42 crc kubenswrapper[4677]: I1203 14:02:42.154670 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-88h22"] Dec 03 14:02:42 crc kubenswrapper[4677]: I1203 14:02:42.154718 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-88h22"] Dec 03 14:02:42 crc kubenswrapper[4677]: I1203 14:02:42.172094 4677 scope.go:117] "RemoveContainer" containerID="87bebbc70b3f6a9a0d8a8cf56c9cadf2991b4ea9b9c4d1d384ee1f9c3fa2b5ee" Dec 03 14:02:42 crc kubenswrapper[4677]: I1203 14:02:42.190899 4677 scope.go:117] "RemoveContainer" containerID="508583c9d979c4fe32fa010f8df19cfd51a9077a7da7472d4a4e65b4f1a52698" Dec 03 14:02:43 crc kubenswrapper[4677]: I1203 14:02:43.983664 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" path="/var/lib/kubelet/pods/36869ae6-cec0-45e4-ae7e-9a703c52ad77/volumes" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964305 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x"] Dec 03 14:02:44 crc kubenswrapper[4677]: E1203 14:02:44.964594 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="extract-content" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964611 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="extract-content" Dec 03 14:02:44 crc kubenswrapper[4677]: E1203 14:02:44.964621 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="registry-server" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964628 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="registry-server" Dec 03 14:02:44 crc kubenswrapper[4677]: E1203 14:02:44.964650 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="registry-server" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964662 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="registry-server" Dec 03 14:02:44 crc kubenswrapper[4677]: E1203 14:02:44.964675 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="extract-utilities" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964682 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="extract-utilities" Dec 03 14:02:44 crc kubenswrapper[4677]: E1203 14:02:44.964694 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="extract-content" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964700 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="extract-content" Dec 03 14:02:44 crc kubenswrapper[4677]: E1203 14:02:44.964725 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="extract-utilities" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964732 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="extract-utilities" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964869 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd86cae-8c5e-4685-b38c-6ce56ed49d1f" containerName="registry-server" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.964895 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="36869ae6-cec0-45e4-ae7e-9a703c52ad77" containerName="registry-server" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.965846 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.971709 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 14:02:44 crc kubenswrapper[4677]: I1203 14:02:44.976144 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x"] Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.164569 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.164621 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.164683 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5jg6\" (UniqueName: \"kubernetes.io/projected/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-kube-api-access-v5jg6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.265830 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.265907 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.266020 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5jg6\" (UniqueName: \"kubernetes.io/projected/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-kube-api-access-v5jg6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.266325 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.266437 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.290290 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5jg6\" (UniqueName: \"kubernetes.io/projected/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-kube-api-access-v5jg6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.585985 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:45 crc kubenswrapper[4677]: I1203 14:02:45.772415 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x"] Dec 03 14:02:46 crc kubenswrapper[4677]: I1203 14:02:46.164189 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" event={"ID":"bbf3592d-dd63-4344-8ae0-2b3b0e88b223","Type":"ContainerStarted","Data":"7c65320631c86aec907a4687c80158c5ab317ff17f399f8d5f1fe6a4b895bcbd"} Dec 03 14:02:47 crc kubenswrapper[4677]: I1203 14:02:47.172460 4677 generic.go:334] "Generic (PLEG): container finished" podID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerID="76844c36be99eb5a9ac85cb1d6b5d8164ba251d449dac751cb0bf4b02ad62f45" exitCode=0 Dec 03 14:02:47 crc kubenswrapper[4677]: I1203 14:02:47.172519 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" event={"ID":"bbf3592d-dd63-4344-8ae0-2b3b0e88b223","Type":"ContainerDied","Data":"76844c36be99eb5a9ac85cb1d6b5d8164ba251d449dac751cb0bf4b02ad62f45"} Dec 03 14:02:49 crc kubenswrapper[4677]: I1203 14:02:49.185852 4677 generic.go:334] "Generic (PLEG): container finished" podID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerID="966f0ba60f83426d3a8da90e724dab0833d72ce46edbe6a56541ebb3fa491627" exitCode=0 Dec 03 14:02:49 crc kubenswrapper[4677]: I1203 14:02:49.185921 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" event={"ID":"bbf3592d-dd63-4344-8ae0-2b3b0e88b223","Type":"ContainerDied","Data":"966f0ba60f83426d3a8da90e724dab0833d72ce46edbe6a56541ebb3fa491627"} Dec 03 14:02:50 crc kubenswrapper[4677]: I1203 14:02:50.193000 4677 generic.go:334] "Generic (PLEG): container finished" podID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerID="081d89d88ae1299279f8e47fbd710d6e9b997a4d1362984202e3aba726a35d59" exitCode=0 Dec 03 14:02:50 crc kubenswrapper[4677]: I1203 14:02:50.193074 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" event={"ID":"bbf3592d-dd63-4344-8ae0-2b3b0e88b223","Type":"ContainerDied","Data":"081d89d88ae1299279f8e47fbd710d6e9b997a4d1362984202e3aba726a35d59"} Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.462476 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.652576 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5jg6\" (UniqueName: \"kubernetes.io/projected/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-kube-api-access-v5jg6\") pod \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.652717 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-util\") pod \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.652773 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-bundle\") pod \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\" (UID: \"bbf3592d-dd63-4344-8ae0-2b3b0e88b223\") " Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.653290 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-bundle" (OuterVolumeSpecName: "bundle") pod "bbf3592d-dd63-4344-8ae0-2b3b0e88b223" (UID: "bbf3592d-dd63-4344-8ae0-2b3b0e88b223"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.657883 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-kube-api-access-v5jg6" (OuterVolumeSpecName: "kube-api-access-v5jg6") pod "bbf3592d-dd63-4344-8ae0-2b3b0e88b223" (UID: "bbf3592d-dd63-4344-8ae0-2b3b0e88b223"). InnerVolumeSpecName "kube-api-access-v5jg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.667698 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-util" (OuterVolumeSpecName: "util") pod "bbf3592d-dd63-4344-8ae0-2b3b0e88b223" (UID: "bbf3592d-dd63-4344-8ae0-2b3b0e88b223"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.732529 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q744q"] Dec 03 14:02:51 crc kubenswrapper[4677]: E1203 14:02:51.732777 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerName="pull" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.732790 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerName="pull" Dec 03 14:02:51 crc kubenswrapper[4677]: E1203 14:02:51.732809 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerName="extract" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.732816 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerName="extract" Dec 03 14:02:51 crc kubenswrapper[4677]: E1203 14:02:51.732830 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerName="util" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.732837 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerName="util" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.732968 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbf3592d-dd63-4344-8ae0-2b3b0e88b223" containerName="extract" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.733762 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.745587 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q744q"] Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.754424 4677 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-util\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.754460 4677 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.754472 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5jg6\" (UniqueName: \"kubernetes.io/projected/bbf3592d-dd63-4344-8ae0-2b3b0e88b223-kube-api-access-v5jg6\") on node \"crc\" DevicePath \"\"" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.855316 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-utilities\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.855371 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-catalog-content\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.855417 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvr7z\" (UniqueName: \"kubernetes.io/projected/6a8ba40b-34d5-49f6-9c81-165c56e006b1-kube-api-access-vvr7z\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.956274 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-utilities\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.956331 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-catalog-content\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.956364 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvr7z\" (UniqueName: \"kubernetes.io/projected/6a8ba40b-34d5-49f6-9c81-165c56e006b1-kube-api-access-vvr7z\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.957066 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-catalog-content\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.957183 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-utilities\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:51 crc kubenswrapper[4677]: I1203 14:02:51.975530 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvr7z\" (UniqueName: \"kubernetes.io/projected/6a8ba40b-34d5-49f6-9c81-165c56e006b1-kube-api-access-vvr7z\") pod \"community-operators-q744q\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:52 crc kubenswrapper[4677]: I1203 14:02:52.069249 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:02:52 crc kubenswrapper[4677]: I1203 14:02:52.211905 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" event={"ID":"bbf3592d-dd63-4344-8ae0-2b3b0e88b223","Type":"ContainerDied","Data":"7c65320631c86aec907a4687c80158c5ab317ff17f399f8d5f1fe6a4b895bcbd"} Dec 03 14:02:52 crc kubenswrapper[4677]: I1203 14:02:52.211966 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c65320631c86aec907a4687c80158c5ab317ff17f399f8d5f1fe6a4b895bcbd" Dec 03 14:02:52 crc kubenswrapper[4677]: I1203 14:02:52.212073 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x" Dec 03 14:02:52 crc kubenswrapper[4677]: I1203 14:02:52.338841 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q744q"] Dec 03 14:02:52 crc kubenswrapper[4677]: W1203 14:02:52.345034 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a8ba40b_34d5_49f6_9c81_165c56e006b1.slice/crio-d7b76db433199dde6fb2e35f33979a2b3e57a4324977078b2e8959f286fa2027 WatchSource:0}: Error finding container d7b76db433199dde6fb2e35f33979a2b3e57a4324977078b2e8959f286fa2027: Status 404 returned error can't find the container with id d7b76db433199dde6fb2e35f33979a2b3e57a4324977078b2e8959f286fa2027 Dec 03 14:02:53 crc kubenswrapper[4677]: I1203 14:02:53.218444 4677 generic.go:334] "Generic (PLEG): container finished" podID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerID="ce4b0ab6d95c7d5ef879f2e92b3e5d60235be4544b47bc159a9744512b8cdcec" exitCode=0 Dec 03 14:02:53 crc kubenswrapper[4677]: I1203 14:02:53.219125 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q744q" event={"ID":"6a8ba40b-34d5-49f6-9c81-165c56e006b1","Type":"ContainerDied","Data":"ce4b0ab6d95c7d5ef879f2e92b3e5d60235be4544b47bc159a9744512b8cdcec"} Dec 03 14:02:53 crc kubenswrapper[4677]: I1203 14:02:53.219215 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q744q" event={"ID":"6a8ba40b-34d5-49f6-9c81-165c56e006b1","Type":"ContainerStarted","Data":"d7b76db433199dde6fb2e35f33979a2b3e57a4324977078b2e8959f286fa2027"} Dec 03 14:02:54 crc kubenswrapper[4677]: I1203 14:02:54.227615 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q744q" event={"ID":"6a8ba40b-34d5-49f6-9c81-165c56e006b1","Type":"ContainerStarted","Data":"39b038d6035c96f38da48eba954d1180fd5bd9b66a9cc8a966c4aefd5aaf306c"} Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.235048 4677 generic.go:334] "Generic (PLEG): container finished" podID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerID="39b038d6035c96f38da48eba954d1180fd5bd9b66a9cc8a966c4aefd5aaf306c" exitCode=0 Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.235119 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q744q" event={"ID":"6a8ba40b-34d5-49f6-9c81-165c56e006b1","Type":"ContainerDied","Data":"39b038d6035c96f38da48eba954d1180fd5bd9b66a9cc8a966c4aefd5aaf306c"} Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.622110 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c"] Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.623223 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.625060 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.625304 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.625522 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-x2mql" Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.644538 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c"] Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.805912 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rw8p\" (UniqueName: \"kubernetes.io/projected/a40313ea-273b-4c40-b414-d492a94efb8b-kube-api-access-7rw8p\") pod \"nmstate-operator-5b5b58f5c8-2x69c\" (UID: \"a40313ea-273b-4c40-b414-d492a94efb8b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.907516 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rw8p\" (UniqueName: \"kubernetes.io/projected/a40313ea-273b-4c40-b414-d492a94efb8b-kube-api-access-7rw8p\") pod \"nmstate-operator-5b5b58f5c8-2x69c\" (UID: \"a40313ea-273b-4c40-b414-d492a94efb8b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.934526 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rw8p\" (UniqueName: \"kubernetes.io/projected/a40313ea-273b-4c40-b414-d492a94efb8b-kube-api-access-7rw8p\") pod \"nmstate-operator-5b5b58f5c8-2x69c\" (UID: \"a40313ea-273b-4c40-b414-d492a94efb8b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" Dec 03 14:02:55 crc kubenswrapper[4677]: I1203 14:02:55.941753 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" Dec 03 14:02:56 crc kubenswrapper[4677]: I1203 14:02:56.244940 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q744q" event={"ID":"6a8ba40b-34d5-49f6-9c81-165c56e006b1","Type":"ContainerStarted","Data":"51b4452d3fe862847c920ffb174771ad398be9f1935b9155d23bb9db30cd34b3"} Dec 03 14:02:56 crc kubenswrapper[4677]: I1203 14:02:56.268758 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q744q" podStartSLOduration=2.837960816 podStartE2EDuration="5.268733529s" podCreationTimestamp="2025-12-03 14:02:51 +0000 UTC" firstStartedPulling="2025-12-03 14:02:53.220484323 +0000 UTC m=+963.966816778" lastFinishedPulling="2025-12-03 14:02:55.651257036 +0000 UTC m=+966.397589491" observedRunningTime="2025-12-03 14:02:56.265796093 +0000 UTC m=+967.012128548" watchObservedRunningTime="2025-12-03 14:02:56.268733529 +0000 UTC m=+967.015065994" Dec 03 14:02:56 crc kubenswrapper[4677]: I1203 14:02:56.355281 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c"] Dec 03 14:02:56 crc kubenswrapper[4677]: W1203 14:02:56.358903 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda40313ea_273b_4c40_b414_d492a94efb8b.slice/crio-d4e97098456c3af7303ca2a23d39eea3b72ed7acac57e1da59db0a382d5e2f89 WatchSource:0}: Error finding container d4e97098456c3af7303ca2a23d39eea3b72ed7acac57e1da59db0a382d5e2f89: Status 404 returned error can't find the container with id d4e97098456c3af7303ca2a23d39eea3b72ed7acac57e1da59db0a382d5e2f89 Dec 03 14:02:57 crc kubenswrapper[4677]: I1203 14:02:57.251308 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" event={"ID":"a40313ea-273b-4c40-b414-d492a94efb8b","Type":"ContainerStarted","Data":"d4e97098456c3af7303ca2a23d39eea3b72ed7acac57e1da59db0a382d5e2f89"} Dec 03 14:02:59 crc kubenswrapper[4677]: I1203 14:02:59.272769 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" event={"ID":"a40313ea-273b-4c40-b414-d492a94efb8b","Type":"ContainerStarted","Data":"356ab148c0b12d2b181292fbaa9912a8f8bbdacdb5b1d683d0ba3ebcc43f8cbe"} Dec 03 14:02:59 crc kubenswrapper[4677]: I1203 14:02:59.288444 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2x69c" podStartSLOduration=1.832935287 podStartE2EDuration="4.288428542s" podCreationTimestamp="2025-12-03 14:02:55 +0000 UTC" firstStartedPulling="2025-12-03 14:02:56.360965539 +0000 UTC m=+967.107297994" lastFinishedPulling="2025-12-03 14:02:58.816458794 +0000 UTC m=+969.562791249" observedRunningTime="2025-12-03 14:02:59.285183517 +0000 UTC m=+970.031515972" watchObservedRunningTime="2025-12-03 14:02:59.288428542 +0000 UTC m=+970.034760997" Dec 03 14:03:02 crc kubenswrapper[4677]: I1203 14:03:02.069880 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:03:02 crc kubenswrapper[4677]: I1203 14:03:02.070362 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:03:02 crc kubenswrapper[4677]: I1203 14:03:02.121358 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:03:02 crc kubenswrapper[4677]: I1203 14:03:02.343343 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:03:03 crc kubenswrapper[4677]: I1203 14:03:03.726614 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q744q"] Dec 03 14:03:05 crc kubenswrapper[4677]: I1203 14:03:05.305093 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q744q" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="registry-server" containerID="cri-o://51b4452d3fe862847c920ffb174771ad398be9f1935b9155d23bb9db30cd34b3" gracePeriod=2 Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.146082 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.147193 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.151074 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-5hhw4" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.159888 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.171815 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-8bpr7"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.172878 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.175206 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.175897 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.178007 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.202118 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.250774 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjstn\" (UniqueName: \"kubernetes.io/projected/24ca8470-a377-4746-a521-8af29247ef95-kube-api-access-bjstn\") pod \"nmstate-metrics-7f946cbc9-ghccd\" (UID: \"24ca8470-a377-4746-a521-8af29247ef95\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.298034 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.299237 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.301176 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.302471 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jpl4d" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.302598 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.311662 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.324392 4677 generic.go:334] "Generic (PLEG): container finished" podID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerID="51b4452d3fe862847c920ffb174771ad398be9f1935b9155d23bb9db30cd34b3" exitCode=0 Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.324450 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q744q" event={"ID":"6a8ba40b-34d5-49f6-9c81-165c56e006b1","Type":"ContainerDied","Data":"51b4452d3fe862847c920ffb174771ad398be9f1935b9155d23bb9db30cd34b3"} Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.351808 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm2m9\" (UniqueName: \"kubernetes.io/projected/80447395-7700-4ff3-9e8c-f18252cf98eb-kube-api-access-vm2m9\") pod \"nmstate-webhook-5f6d4c5ccb-zlfwn\" (UID: \"80447395-7700-4ff3-9e8c-f18252cf98eb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.351868 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-dbus-socket\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.351899 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-ovs-socket\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.351938 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/80447395-7700-4ff3-9e8c-f18252cf98eb-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-zlfwn\" (UID: \"80447395-7700-4ff3-9e8c-f18252cf98eb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.351987 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxbhl\" (UniqueName: \"kubernetes.io/projected/feb427e2-dfd2-4de1-9f57-e230be83a451-kube-api-access-fxbhl\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.352038 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-nmstate-lock\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.352073 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjstn\" (UniqueName: \"kubernetes.io/projected/24ca8470-a377-4746-a521-8af29247ef95-kube-api-access-bjstn\") pod \"nmstate-metrics-7f946cbc9-ghccd\" (UID: \"24ca8470-a377-4746-a521-8af29247ef95\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.376844 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjstn\" (UniqueName: \"kubernetes.io/projected/24ca8470-a377-4746-a521-8af29247ef95-kube-api-access-bjstn\") pod \"nmstate-metrics-7f946cbc9-ghccd\" (UID: \"24ca8470-a377-4746-a521-8af29247ef95\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453527 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-dbus-socket\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453590 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-ovs-socket\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453618 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b745bd9-ab3c-42d4-9476-5460897f8164-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453649 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/80447395-7700-4ff3-9e8c-f18252cf98eb-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-zlfwn\" (UID: \"80447395-7700-4ff3-9e8c-f18252cf98eb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453668 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4b745bd9-ab3c-42d4-9476-5460897f8164-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453690 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8bw6\" (UniqueName: \"kubernetes.io/projected/4b745bd9-ab3c-42d4-9476-5460897f8164-kube-api-access-j8bw6\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453761 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-ovs-socket\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: E1203 14:03:06.453791 4677 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 03 14:03:06 crc kubenswrapper[4677]: E1203 14:03:06.453857 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/80447395-7700-4ff3-9e8c-f18252cf98eb-tls-key-pair podName:80447395-7700-4ff3-9e8c-f18252cf98eb nodeName:}" failed. No retries permitted until 2025-12-03 14:03:06.953834503 +0000 UTC m=+977.700166968 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/80447395-7700-4ff3-9e8c-f18252cf98eb-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-zlfwn" (UID: "80447395-7700-4ff3-9e8c-f18252cf98eb") : secret "openshift-nmstate-webhook" not found Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453877 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxbhl\" (UniqueName: \"kubernetes.io/projected/feb427e2-dfd2-4de1-9f57-e230be83a451-kube-api-access-fxbhl\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453897 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-dbus-socket\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453967 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-nmstate-lock\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.453922 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/feb427e2-dfd2-4de1-9f57-e230be83a451-nmstate-lock\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.454096 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm2m9\" (UniqueName: \"kubernetes.io/projected/80447395-7700-4ff3-9e8c-f18252cf98eb-kube-api-access-vm2m9\") pod \"nmstate-webhook-5f6d4c5ccb-zlfwn\" (UID: \"80447395-7700-4ff3-9e8c-f18252cf98eb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.474898 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm2m9\" (UniqueName: \"kubernetes.io/projected/80447395-7700-4ff3-9e8c-f18252cf98eb-kube-api-access-vm2m9\") pod \"nmstate-webhook-5f6d4c5ccb-zlfwn\" (UID: \"80447395-7700-4ff3-9e8c-f18252cf98eb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.477615 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxbhl\" (UniqueName: \"kubernetes.io/projected/feb427e2-dfd2-4de1-9f57-e230be83a451-kube-api-access-fxbhl\") pod \"nmstate-handler-8bpr7\" (UID: \"feb427e2-dfd2-4de1-9f57-e230be83a451\") " pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.479327 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.521184 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.531508 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6d4575bd56-c4z57"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.532462 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.555248 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8bw6\" (UniqueName: \"kubernetes.io/projected/4b745bd9-ab3c-42d4-9476-5460897f8164-kube-api-access-j8bw6\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.555354 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b745bd9-ab3c-42d4-9476-5460897f8164-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.555366 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d4575bd56-c4z57"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.555413 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4b745bd9-ab3c-42d4-9476-5460897f8164-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.556576 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4b745bd9-ab3c-42d4-9476-5460897f8164-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: E1203 14:03:06.558129 4677 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 03 14:03:06 crc kubenswrapper[4677]: E1203 14:03:06.558218 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b745bd9-ab3c-42d4-9476-5460897f8164-plugin-serving-cert podName:4b745bd9-ab3c-42d4-9476-5460897f8164 nodeName:}" failed. No retries permitted until 2025-12-03 14:03:07.058200928 +0000 UTC m=+977.804533383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/4b745bd9-ab3c-42d4-9476-5460897f8164-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-5rflb" (UID: "4b745bd9-ab3c-42d4-9476-5460897f8164") : secret "plugin-serving-cert" not found Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.587479 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8bw6\" (UniqueName: \"kubernetes.io/projected/4b745bd9-ab3c-42d4-9476-5460897f8164-kube-api-access-j8bw6\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.656774 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad36d4b-b953-479e-ad36-97665e16f672-console-serving-cert\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.656857 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p48z4\" (UniqueName: \"kubernetes.io/projected/9ad36d4b-b953-479e-ad36-97665e16f672-kube-api-access-p48z4\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.657187 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9ad36d4b-b953-479e-ad36-97665e16f672-console-oauth-config\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.657223 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-trusted-ca-bundle\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.657257 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-service-ca\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.657283 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-oauth-serving-cert\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.657311 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-console-config\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.747127 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd"] Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.757984 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-service-ca\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.758014 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-oauth-serving-cert\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.758030 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-console-config\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.758089 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad36d4b-b953-479e-ad36-97665e16f672-console-serving-cert\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.758140 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p48z4\" (UniqueName: \"kubernetes.io/projected/9ad36d4b-b953-479e-ad36-97665e16f672-kube-api-access-p48z4\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.758174 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9ad36d4b-b953-479e-ad36-97665e16f672-console-oauth-config\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.758211 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-trusted-ca-bundle\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.758862 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-service-ca\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.759231 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-trusted-ca-bundle\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.759881 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-oauth-serving-cert\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.760497 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9ad36d4b-b953-479e-ad36-97665e16f672-console-config\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.762702 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad36d4b-b953-479e-ad36-97665e16f672-console-serving-cert\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.762916 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9ad36d4b-b953-479e-ad36-97665e16f672-console-oauth-config\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.791493 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p48z4\" (UniqueName: \"kubernetes.io/projected/9ad36d4b-b953-479e-ad36-97665e16f672-kube-api-access-p48z4\") pod \"console-6d4575bd56-c4z57\" (UID: \"9ad36d4b-b953-479e-ad36-97665e16f672\") " pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.920612 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.967692 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/80447395-7700-4ff3-9e8c-f18252cf98eb-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-zlfwn\" (UID: \"80447395-7700-4ff3-9e8c-f18252cf98eb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:06 crc kubenswrapper[4677]: I1203 14:03:06.971580 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/80447395-7700-4ff3-9e8c-f18252cf98eb-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-zlfwn\" (UID: \"80447395-7700-4ff3-9e8c-f18252cf98eb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.069916 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b745bd9-ab3c-42d4-9476-5460897f8164-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.078679 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b745bd9-ab3c-42d4-9476-5460897f8164-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5rflb\" (UID: \"4b745bd9-ab3c-42d4-9476-5460897f8164\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.125193 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.130012 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:07 crc kubenswrapper[4677]: W1203 14:03:07.179750 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ad36d4b_b953_479e_ad36_97665e16f672.slice/crio-77a90952165deb8bab3e89b9d462058acc77fc21138742e3808401ad2991d9f8 WatchSource:0}: Error finding container 77a90952165deb8bab3e89b9d462058acc77fc21138742e3808401ad2991d9f8: Status 404 returned error can't find the container with id 77a90952165deb8bab3e89b9d462058acc77fc21138742e3808401ad2991d9f8 Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.183240 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d4575bd56-c4z57"] Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.222861 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.275247 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvr7z\" (UniqueName: \"kubernetes.io/projected/6a8ba40b-34d5-49f6-9c81-165c56e006b1-kube-api-access-vvr7z\") pod \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.275679 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-catalog-content\") pod \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.275772 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-utilities\") pod \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\" (UID: \"6a8ba40b-34d5-49f6-9c81-165c56e006b1\") " Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.277932 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-utilities" (OuterVolumeSpecName: "utilities") pod "6a8ba40b-34d5-49f6-9c81-165c56e006b1" (UID: "6a8ba40b-34d5-49f6-9c81-165c56e006b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.279673 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8ba40b-34d5-49f6-9c81-165c56e006b1-kube-api-access-vvr7z" (OuterVolumeSpecName: "kube-api-access-vvr7z") pod "6a8ba40b-34d5-49f6-9c81-165c56e006b1" (UID: "6a8ba40b-34d5-49f6-9c81-165c56e006b1"). InnerVolumeSpecName "kube-api-access-vvr7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.332018 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" event={"ID":"24ca8470-a377-4746-a521-8af29247ef95","Type":"ContainerStarted","Data":"d5c26f979c375f5b1d35488358fbbea7302845b41e3b9f9f4f0d8bba7dbcaaca"} Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.333747 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q744q" event={"ID":"6a8ba40b-34d5-49f6-9c81-165c56e006b1","Type":"ContainerDied","Data":"d7b76db433199dde6fb2e35f33979a2b3e57a4324977078b2e8959f286fa2027"} Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.333805 4677 scope.go:117] "RemoveContainer" containerID="51b4452d3fe862847c920ffb174771ad398be9f1935b9155d23bb9db30cd34b3" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.333920 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q744q" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.337726 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-8bpr7" event={"ID":"feb427e2-dfd2-4de1-9f57-e230be83a451","Type":"ContainerStarted","Data":"b0a936271a61cfc5bb79480b1ab0d988693b2db29ccc6fbcfa3f80de49dcdd3d"} Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.341638 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d4575bd56-c4z57" event={"ID":"9ad36d4b-b953-479e-ad36-97665e16f672","Type":"ContainerStarted","Data":"77a90952165deb8bab3e89b9d462058acc77fc21138742e3808401ad2991d9f8"} Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.351467 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a8ba40b-34d5-49f6-9c81-165c56e006b1" (UID: "6a8ba40b-34d5-49f6-9c81-165c56e006b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.358063 4677 scope.go:117] "RemoveContainer" containerID="39b038d6035c96f38da48eba954d1180fd5bd9b66a9cc8a966c4aefd5aaf306c" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.363357 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn"] Dec 03 14:03:07 crc kubenswrapper[4677]: W1203 14:03:07.369357 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80447395_7700_4ff3_9e8c_f18252cf98eb.slice/crio-f28f72dc7a8ceec4768b26b4aac046f7daf3d1e1ea371f44c75ce55a950fd6a8 WatchSource:0}: Error finding container f28f72dc7a8ceec4768b26b4aac046f7daf3d1e1ea371f44c75ce55a950fd6a8: Status 404 returned error can't find the container with id f28f72dc7a8ceec4768b26b4aac046f7daf3d1e1ea371f44c75ce55a950fd6a8 Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.379305 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvr7z\" (UniqueName: \"kubernetes.io/projected/6a8ba40b-34d5-49f6-9c81-165c56e006b1-kube-api-access-vvr7z\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.379336 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.379349 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ba40b-34d5-49f6-9c81-165c56e006b1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.384881 4677 scope.go:117] "RemoveContainer" containerID="ce4b0ab6d95c7d5ef879f2e92b3e5d60235be4544b47bc159a9744512b8cdcec" Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.434007 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb"] Dec 03 14:03:07 crc kubenswrapper[4677]: W1203 14:03:07.441117 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b745bd9_ab3c_42d4_9476_5460897f8164.slice/crio-28f561151c0c42f0d2a1b50e0cfb037421437a26600f1d3f905985deb999ceb1 WatchSource:0}: Error finding container 28f561151c0c42f0d2a1b50e0cfb037421437a26600f1d3f905985deb999ceb1: Status 404 returned error can't find the container with id 28f561151c0c42f0d2a1b50e0cfb037421437a26600f1d3f905985deb999ceb1 Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.662165 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q744q"] Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.666098 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q744q"] Dec 03 14:03:07 crc kubenswrapper[4677]: I1203 14:03:07.984063 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" path="/var/lib/kubelet/pods/6a8ba40b-34d5-49f6-9c81-165c56e006b1/volumes" Dec 03 14:03:08 crc kubenswrapper[4677]: I1203 14:03:08.348662 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" event={"ID":"4b745bd9-ab3c-42d4-9476-5460897f8164","Type":"ContainerStarted","Data":"28f561151c0c42f0d2a1b50e0cfb037421437a26600f1d3f905985deb999ceb1"} Dec 03 14:03:08 crc kubenswrapper[4677]: I1203 14:03:08.349682 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" event={"ID":"80447395-7700-4ff3-9e8c-f18252cf98eb","Type":"ContainerStarted","Data":"f28f72dc7a8ceec4768b26b4aac046f7daf3d1e1ea371f44c75ce55a950fd6a8"} Dec 03 14:03:08 crc kubenswrapper[4677]: I1203 14:03:08.351852 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d4575bd56-c4z57" event={"ID":"9ad36d4b-b953-479e-ad36-97665e16f672","Type":"ContainerStarted","Data":"2d6779df1722041efb58bb96871a4241712044adde7897180aa27c06ec28b837"} Dec 03 14:03:08 crc kubenswrapper[4677]: I1203 14:03:08.369470 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6d4575bd56-c4z57" podStartSLOduration=2.369454971 podStartE2EDuration="2.369454971s" podCreationTimestamp="2025-12-03 14:03:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:03:08.367704701 +0000 UTC m=+979.114037176" watchObservedRunningTime="2025-12-03 14:03:08.369454971 +0000 UTC m=+979.115787426" Dec 03 14:03:08 crc kubenswrapper[4677]: I1203 14:03:08.437109 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:03:08 crc kubenswrapper[4677]: I1203 14:03:08.437176 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.371145 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-8bpr7" event={"ID":"feb427e2-dfd2-4de1-9f57-e230be83a451","Type":"ContainerStarted","Data":"bf433c384305455f01dfef391a0a630d70dbde5d5712e819f5a20f6550984030"} Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.371739 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.381668 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" event={"ID":"4b745bd9-ab3c-42d4-9476-5460897f8164","Type":"ContainerStarted","Data":"cc320397f13e891a7e80a76192407c921af97c52b8895a96fb0cafb40ac430c5"} Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.412494 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-8bpr7" podStartSLOduration=1.220352293 podStartE2EDuration="5.412478086s" podCreationTimestamp="2025-12-03 14:03:06 +0000 UTC" firstStartedPulling="2025-12-03 14:03:06.585239036 +0000 UTC m=+977.331571491" lastFinishedPulling="2025-12-03 14:03:10.777364829 +0000 UTC m=+981.523697284" observedRunningTime="2025-12-03 14:03:11.410373444 +0000 UTC m=+982.156705899" watchObservedRunningTime="2025-12-03 14:03:11.412478086 +0000 UTC m=+982.158810541" Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.415212 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" event={"ID":"24ca8470-a377-4746-a521-8af29247ef95","Type":"ContainerStarted","Data":"222dc0f1968c000c6d8b89417f32b9bcb8cf9340a3a7924c92cddd558238e51c"} Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.419683 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" event={"ID":"80447395-7700-4ff3-9e8c-f18252cf98eb","Type":"ContainerStarted","Data":"9c77681e2294c981d18f163fd35316000b4f2007dd27149f8a71eba522e5cfcb"} Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.420006 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.427129 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5rflb" podStartSLOduration=2.019508264 podStartE2EDuration="5.427113642s" podCreationTimestamp="2025-12-03 14:03:06 +0000 UTC" firstStartedPulling="2025-12-03 14:03:07.443598935 +0000 UTC m=+978.189931390" lastFinishedPulling="2025-12-03 14:03:10.851204313 +0000 UTC m=+981.597536768" observedRunningTime="2025-12-03 14:03:11.426565356 +0000 UTC m=+982.172897811" watchObservedRunningTime="2025-12-03 14:03:11.427113642 +0000 UTC m=+982.173446097" Dec 03 14:03:11 crc kubenswrapper[4677]: I1203 14:03:11.449782 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" podStartSLOduration=1.986335666 podStartE2EDuration="5.449768563s" podCreationTimestamp="2025-12-03 14:03:06 +0000 UTC" firstStartedPulling="2025-12-03 14:03:07.384708127 +0000 UTC m=+978.131040582" lastFinishedPulling="2025-12-03 14:03:10.848141024 +0000 UTC m=+981.594473479" observedRunningTime="2025-12-03 14:03:11.446398145 +0000 UTC m=+982.192730610" watchObservedRunningTime="2025-12-03 14:03:11.449768563 +0000 UTC m=+982.196101018" Dec 03 14:03:15 crc kubenswrapper[4677]: I1203 14:03:15.456648 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" event={"ID":"24ca8470-a377-4746-a521-8af29247ef95","Type":"ContainerStarted","Data":"94e8412844e0c609940aac7ea7eee9b2dd357a0b326525002dbca50bc32c04ef"} Dec 03 14:03:15 crc kubenswrapper[4677]: I1203 14:03:15.487651 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-ghccd" podStartSLOduration=1.4178200429999999 podStartE2EDuration="9.487631136s" podCreationTimestamp="2025-12-03 14:03:06 +0000 UTC" firstStartedPulling="2025-12-03 14:03:06.757008577 +0000 UTC m=+977.503341032" lastFinishedPulling="2025-12-03 14:03:14.82681967 +0000 UTC m=+985.573152125" observedRunningTime="2025-12-03 14:03:15.484177845 +0000 UTC m=+986.230510310" watchObservedRunningTime="2025-12-03 14:03:15.487631136 +0000 UTC m=+986.233963601" Dec 03 14:03:16 crc kubenswrapper[4677]: I1203 14:03:16.641479 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-8bpr7" Dec 03 14:03:16 crc kubenswrapper[4677]: I1203 14:03:16.921518 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:16 crc kubenswrapper[4677]: I1203 14:03:16.921574 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:16 crc kubenswrapper[4677]: I1203 14:03:16.926816 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:17 crc kubenswrapper[4677]: I1203 14:03:17.476014 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6d4575bd56-c4z57" Dec 03 14:03:17 crc kubenswrapper[4677]: I1203 14:03:17.524735 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s84pq"] Dec 03 14:03:27 crc kubenswrapper[4677]: I1203 14:03:27.136905 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-zlfwn" Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.437761 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.438660 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.438717 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.439456 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1186ac74eab5d1036a6f0e4e1fb80c6c8725f92e6867eb61894f1292d51d689c"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.439531 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://1186ac74eab5d1036a6f0e4e1fb80c6c8725f92e6867eb61894f1292d51d689c" gracePeriod=600 Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.617355 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="1186ac74eab5d1036a6f0e4e1fb80c6c8725f92e6867eb61894f1292d51d689c" exitCode=0 Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.617713 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"1186ac74eab5d1036a6f0e4e1fb80c6c8725f92e6867eb61894f1292d51d689c"} Dec 03 14:03:38 crc kubenswrapper[4677]: I1203 14:03:38.617750 4677 scope.go:117] "RemoveContainer" containerID="66093cdea49042d1d130e4c423008a09d5d1d6840fa98fa247eeb0b6e0fa1441" Dec 03 14:03:39 crc kubenswrapper[4677]: I1203 14:03:39.624763 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"2e1aec0429580d81a4c5a21902c4598063a10937c54a8f7496939174766a6f58"} Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.844832 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp"] Dec 03 14:03:40 crc kubenswrapper[4677]: E1203 14:03:40.845345 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="extract-utilities" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.845357 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="extract-utilities" Dec 03 14:03:40 crc kubenswrapper[4677]: E1203 14:03:40.845373 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="extract-content" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.845379 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="extract-content" Dec 03 14:03:40 crc kubenswrapper[4677]: E1203 14:03:40.845390 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="registry-server" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.845397 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="registry-server" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.845497 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8ba40b-34d5-49f6-9c81-165c56e006b1" containerName="registry-server" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.846312 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.849998 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.864016 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp"] Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.952794 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.952971 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:40 crc kubenswrapper[4677]: I1203 14:03:40.953000 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xjdb\" (UniqueName: \"kubernetes.io/projected/ef1bb5ce-06d7-47db-a43c-85e57f070cff-kube-api-access-6xjdb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.054510 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.054550 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xjdb\" (UniqueName: \"kubernetes.io/projected/ef1bb5ce-06d7-47db-a43c-85e57f070cff-kube-api-access-6xjdb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.054576 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.055035 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.055051 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.080728 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xjdb\" (UniqueName: \"kubernetes.io/projected/ef1bb5ce-06d7-47db-a43c-85e57f070cff-kube-api-access-6xjdb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.164340 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.414617 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp"] Dec 03 14:03:41 crc kubenswrapper[4677]: W1203 14:03:41.426331 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef1bb5ce_06d7_47db_a43c_85e57f070cff.slice/crio-e8aad544fa3d8c0db842816d3cbc51c7f2fa17a121d919ab7a084b3fbbb11cf5 WatchSource:0}: Error finding container e8aad544fa3d8c0db842816d3cbc51c7f2fa17a121d919ab7a084b3fbbb11cf5: Status 404 returned error can't find the container with id e8aad544fa3d8c0db842816d3cbc51c7f2fa17a121d919ab7a084b3fbbb11cf5 Dec 03 14:03:41 crc kubenswrapper[4677]: I1203 14:03:41.643648 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" event={"ID":"ef1bb5ce-06d7-47db-a43c-85e57f070cff","Type":"ContainerStarted","Data":"e8aad544fa3d8c0db842816d3cbc51c7f2fa17a121d919ab7a084b3fbbb11cf5"} Dec 03 14:03:42 crc kubenswrapper[4677]: I1203 14:03:42.570427 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-s84pq" podUID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" containerName="console" containerID="cri-o://dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de" gracePeriod=15 Dec 03 14:03:42 crc kubenswrapper[4677]: I1203 14:03:42.651219 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerID="5811db2aefd3e203cde5a4d2fda52f74cbff98212c14918accf3370d4d28d0c8" exitCode=0 Dec 03 14:03:42 crc kubenswrapper[4677]: I1203 14:03:42.651272 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" event={"ID":"ef1bb5ce-06d7-47db-a43c-85e57f070cff","Type":"ContainerDied","Data":"5811db2aefd3e203cde5a4d2fda52f74cbff98212c14918accf3370d4d28d0c8"} Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.085913 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s84pq_58f5bde4-0a29-4a20-a332-e945cc8bc3a0/console/0.log" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.086293 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.179802 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqzt7\" (UniqueName: \"kubernetes.io/projected/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-kube-api-access-gqzt7\") pod \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.179898 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-serving-cert\") pod \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.180009 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-oauth-serving-cert\") pod \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.180046 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-service-ca\") pod \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.180142 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-config\") pod \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.180184 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-oauth-config\") pod \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.180207 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-trusted-ca-bundle\") pod \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\" (UID: \"58f5bde4-0a29-4a20-a332-e945cc8bc3a0\") " Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.180720 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "58f5bde4-0a29-4a20-a332-e945cc8bc3a0" (UID: "58f5bde4-0a29-4a20-a332-e945cc8bc3a0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.180866 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-service-ca" (OuterVolumeSpecName: "service-ca") pod "58f5bde4-0a29-4a20-a332-e945cc8bc3a0" (UID: "58f5bde4-0a29-4a20-a332-e945cc8bc3a0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.181397 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-config" (OuterVolumeSpecName: "console-config") pod "58f5bde4-0a29-4a20-a332-e945cc8bc3a0" (UID: "58f5bde4-0a29-4a20-a332-e945cc8bc3a0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.181446 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "58f5bde4-0a29-4a20-a332-e945cc8bc3a0" (UID: "58f5bde4-0a29-4a20-a332-e945cc8bc3a0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.186320 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "58f5bde4-0a29-4a20-a332-e945cc8bc3a0" (UID: "58f5bde4-0a29-4a20-a332-e945cc8bc3a0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.186357 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-kube-api-access-gqzt7" (OuterVolumeSpecName: "kube-api-access-gqzt7") pod "58f5bde4-0a29-4a20-a332-e945cc8bc3a0" (UID: "58f5bde4-0a29-4a20-a332-e945cc8bc3a0"). InnerVolumeSpecName "kube-api-access-gqzt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.187303 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "58f5bde4-0a29-4a20-a332-e945cc8bc3a0" (UID: "58f5bde4-0a29-4a20-a332-e945cc8bc3a0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.282201 4677 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.282252 4677 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.282264 4677 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.282274 4677 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.282285 4677 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.282297 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqzt7\" (UniqueName: \"kubernetes.io/projected/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-kube-api-access-gqzt7\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.282311 4677 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58f5bde4-0a29-4a20-a332-e945cc8bc3a0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.659332 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s84pq_58f5bde4-0a29-4a20-a332-e945cc8bc3a0/console/0.log" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.659632 4677 generic.go:334] "Generic (PLEG): container finished" podID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" containerID="dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de" exitCode=2 Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.659666 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s84pq" event={"ID":"58f5bde4-0a29-4a20-a332-e945cc8bc3a0","Type":"ContainerDied","Data":"dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de"} Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.659698 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s84pq" event={"ID":"58f5bde4-0a29-4a20-a332-e945cc8bc3a0","Type":"ContainerDied","Data":"15b99d25c79a99f2acbd9ffd50d151e9d57bda6f611ca50c834f854ab0c8750a"} Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.659752 4677 scope.go:117] "RemoveContainer" containerID="dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.659751 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s84pq" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.675159 4677 scope.go:117] "RemoveContainer" containerID="dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de" Dec 03 14:03:43 crc kubenswrapper[4677]: E1203 14:03:43.675557 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de\": container with ID starting with dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de not found: ID does not exist" containerID="dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.675590 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de"} err="failed to get container status \"dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de\": rpc error: code = NotFound desc = could not find container \"dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de\": container with ID starting with dc0365a34861b0e0c3125ec713e64a4d6ade0eac14d21a81b98f9f729e03a2de not found: ID does not exist" Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.689289 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s84pq"] Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.694875 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-s84pq"] Dec 03 14:03:43 crc kubenswrapper[4677]: I1203 14:03:43.986443 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" path="/var/lib/kubelet/pods/58f5bde4-0a29-4a20-a332-e945cc8bc3a0/volumes" Dec 03 14:03:47 crc kubenswrapper[4677]: I1203 14:03:47.690456 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerID="418724c4e1b66120d9976dabf1f67e8849fb19b3660bbe0ff4a38bdbf796ba18" exitCode=0 Dec 03 14:03:47 crc kubenswrapper[4677]: I1203 14:03:47.690603 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" event={"ID":"ef1bb5ce-06d7-47db-a43c-85e57f070cff","Type":"ContainerDied","Data":"418724c4e1b66120d9976dabf1f67e8849fb19b3660bbe0ff4a38bdbf796ba18"} Dec 03 14:03:48 crc kubenswrapper[4677]: I1203 14:03:48.701593 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerID="7aa16474b1a646adba7b357c648d978f604c4e09316f731cb163643a2058b716" exitCode=0 Dec 03 14:03:48 crc kubenswrapper[4677]: I1203 14:03:48.701708 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" event={"ID":"ef1bb5ce-06d7-47db-a43c-85e57f070cff","Type":"ContainerDied","Data":"7aa16474b1a646adba7b357c648d978f604c4e09316f731cb163643a2058b716"} Dec 03 14:03:49 crc kubenswrapper[4677]: I1203 14:03:49.951539 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.077328 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-bundle\") pod \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.077720 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xjdb\" (UniqueName: \"kubernetes.io/projected/ef1bb5ce-06d7-47db-a43c-85e57f070cff-kube-api-access-6xjdb\") pod \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.077745 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-util\") pod \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\" (UID: \"ef1bb5ce-06d7-47db-a43c-85e57f070cff\") " Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.078707 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-bundle" (OuterVolumeSpecName: "bundle") pod "ef1bb5ce-06d7-47db-a43c-85e57f070cff" (UID: "ef1bb5ce-06d7-47db-a43c-85e57f070cff"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.082593 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef1bb5ce-06d7-47db-a43c-85e57f070cff-kube-api-access-6xjdb" (OuterVolumeSpecName: "kube-api-access-6xjdb") pod "ef1bb5ce-06d7-47db-a43c-85e57f070cff" (UID: "ef1bb5ce-06d7-47db-a43c-85e57f070cff"). InnerVolumeSpecName "kube-api-access-6xjdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.090452 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-util" (OuterVolumeSpecName: "util") pod "ef1bb5ce-06d7-47db-a43c-85e57f070cff" (UID: "ef1bb5ce-06d7-47db-a43c-85e57f070cff"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.179254 4677 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.179305 4677 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef1bb5ce-06d7-47db-a43c-85e57f070cff-util\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.179325 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xjdb\" (UniqueName: \"kubernetes.io/projected/ef1bb5ce-06d7-47db-a43c-85e57f070cff-kube-api-access-6xjdb\") on node \"crc\" DevicePath \"\"" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.714539 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" event={"ID":"ef1bb5ce-06d7-47db-a43c-85e57f070cff","Type":"ContainerDied","Data":"e8aad544fa3d8c0db842816d3cbc51c7f2fa17a121d919ab7a084b3fbbb11cf5"} Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.714604 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8aad544fa3d8c0db842816d3cbc51c7f2fa17a121d919ab7a084b3fbbb11cf5" Dec 03 14:03:50 crc kubenswrapper[4677]: I1203 14:03:50.714659 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.012752 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r"] Dec 03 14:04:04 crc kubenswrapper[4677]: E1203 14:04:04.015997 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerName="extract" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.016373 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerName="extract" Dec 03 14:04:04 crc kubenswrapper[4677]: E1203 14:04:04.016460 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" containerName="console" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.016532 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" containerName="console" Dec 03 14:04:04 crc kubenswrapper[4677]: E1203 14:04:04.016605 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerName="util" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.016672 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerName="util" Dec 03 14:04:04 crc kubenswrapper[4677]: E1203 14:04:04.016743 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerName="pull" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.016809 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerName="pull" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.017006 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1bb5ce-06d7-47db-a43c-85e57f070cff" containerName="extract" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.017091 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="58f5bde4-0a29-4a20-a332-e945cc8bc3a0" containerName="console" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.017705 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.020183 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.020748 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.020899 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.021274 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.025665 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-xqcqx" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.028538 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r"] Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.070843 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwwxj\" (UniqueName: \"kubernetes.io/projected/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-kube-api-access-cwwxj\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.071143 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-apiservice-cert\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.071271 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-webhook-cert\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.172520 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-apiservice-cert\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.172615 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-webhook-cert\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.172652 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwwxj\" (UniqueName: \"kubernetes.io/projected/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-kube-api-access-cwwxj\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.180280 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-apiservice-cert\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.180967 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-webhook-cert\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.192611 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwwxj\" (UniqueName: \"kubernetes.io/projected/f6eaef50-3b3e-4fb9-a22a-7db928062fa3-kube-api-access-cwwxj\") pod \"metallb-operator-controller-manager-5f48b5ddf5-58d4r\" (UID: \"f6eaef50-3b3e-4fb9-a22a-7db928062fa3\") " pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.335044 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.460335 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c"] Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.461400 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.467169 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-7mfds" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.467363 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.467504 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.479450 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c"] Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.584877 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9427b27d-5375-4914-9481-fbbd6c1f9e21-webhook-cert\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.584923 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9427b27d-5375-4914-9481-fbbd6c1f9e21-apiservice-cert\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.584976 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghpvp\" (UniqueName: \"kubernetes.io/projected/9427b27d-5375-4914-9481-fbbd6c1f9e21-kube-api-access-ghpvp\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.687588 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9427b27d-5375-4914-9481-fbbd6c1f9e21-webhook-cert\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.687668 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9427b27d-5375-4914-9481-fbbd6c1f9e21-apiservice-cert\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.687714 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghpvp\" (UniqueName: \"kubernetes.io/projected/9427b27d-5375-4914-9481-fbbd6c1f9e21-kube-api-access-ghpvp\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.700519 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9427b27d-5375-4914-9481-fbbd6c1f9e21-webhook-cert\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.716755 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9427b27d-5375-4914-9481-fbbd6c1f9e21-apiservice-cert\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.717739 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghpvp\" (UniqueName: \"kubernetes.io/projected/9427b27d-5375-4914-9481-fbbd6c1f9e21-kube-api-access-ghpvp\") pod \"metallb-operator-webhook-server-5dcf4b8dc4-2f58c\" (UID: \"9427b27d-5375-4914-9481-fbbd6c1f9e21\") " pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.781104 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:04 crc kubenswrapper[4677]: I1203 14:04:04.990934 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r"] Dec 03 14:04:05 crc kubenswrapper[4677]: I1203 14:04:05.324636 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c"] Dec 03 14:04:05 crc kubenswrapper[4677]: W1203 14:04:05.335279 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9427b27d_5375_4914_9481_fbbd6c1f9e21.slice/crio-26607a185c9e0e780c90306a6f605d476a90d08a837310b95948809781126cf3 WatchSource:0}: Error finding container 26607a185c9e0e780c90306a6f605d476a90d08a837310b95948809781126cf3: Status 404 returned error can't find the container with id 26607a185c9e0e780c90306a6f605d476a90d08a837310b95948809781126cf3 Dec 03 14:04:05 crc kubenswrapper[4677]: I1203 14:04:05.821714 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" event={"ID":"f6eaef50-3b3e-4fb9-a22a-7db928062fa3","Type":"ContainerStarted","Data":"c4e4cde99a9b3de016bdb3e392247b82825d16f892bfaad1c6ac677f49dd68d0"} Dec 03 14:04:05 crc kubenswrapper[4677]: I1203 14:04:05.822639 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" event={"ID":"9427b27d-5375-4914-9481-fbbd6c1f9e21","Type":"ContainerStarted","Data":"26607a185c9e0e780c90306a6f605d476a90d08a837310b95948809781126cf3"} Dec 03 14:04:13 crc kubenswrapper[4677]: I1203 14:04:13.875880 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" event={"ID":"9427b27d-5375-4914-9481-fbbd6c1f9e21","Type":"ContainerStarted","Data":"4fd68f4a314779acab520c1faaace305bb05080b61057ad4f642be570209636e"} Dec 03 14:04:13 crc kubenswrapper[4677]: I1203 14:04:13.877805 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" event={"ID":"f6eaef50-3b3e-4fb9-a22a-7db928062fa3","Type":"ContainerStarted","Data":"c14473abd96e8132cf374a9890917c6dcd251941767d1768fb28fb26fd6b53cd"} Dec 03 14:04:13 crc kubenswrapper[4677]: I1203 14:04:13.878750 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:14 crc kubenswrapper[4677]: I1203 14:04:14.896537 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:14 crc kubenswrapper[4677]: I1203 14:04:14.919688 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" podStartSLOduration=3.58956751 podStartE2EDuration="10.9196681s" podCreationTimestamp="2025-12-03 14:04:04 +0000 UTC" firstStartedPulling="2025-12-03 14:04:05.33860685 +0000 UTC m=+1036.084939305" lastFinishedPulling="2025-12-03 14:04:12.66870744 +0000 UTC m=+1043.415039895" observedRunningTime="2025-12-03 14:04:14.917001792 +0000 UTC m=+1045.663334267" watchObservedRunningTime="2025-12-03 14:04:14.9196681 +0000 UTC m=+1045.666000565" Dec 03 14:04:14 crc kubenswrapper[4677]: I1203 14:04:14.921961 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" podStartSLOduration=4.266652894 podStartE2EDuration="11.921905748s" podCreationTimestamp="2025-12-03 14:04:03 +0000 UTC" firstStartedPulling="2025-12-03 14:04:04.999349474 +0000 UTC m=+1035.745681929" lastFinishedPulling="2025-12-03 14:04:12.654602328 +0000 UTC m=+1043.400934783" observedRunningTime="2025-12-03 14:04:13.902185683 +0000 UTC m=+1044.648518158" watchObservedRunningTime="2025-12-03 14:04:14.921905748 +0000 UTC m=+1045.668238203" Dec 03 14:04:24 crc kubenswrapper[4677]: I1203 14:04:24.791116 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" Dec 03 14:04:44 crc kubenswrapper[4677]: I1203 14:04:44.339047 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.210970 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-lcgmk"] Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.213155 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.214581 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.217233 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.217356 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dr8zx" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.222755 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q"] Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.224204 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.225528 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.230800 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q"] Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.275866 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/968d4f67-d25c-48f4-b120-a345dc551e55-metrics-certs\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.275919 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnp5x\" (UniqueName: \"kubernetes.io/projected/968d4f67-d25c-48f4-b120-a345dc551e55-kube-api-access-pnp5x\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.275943 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-frr-sockets\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.275990 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pkrd\" (UniqueName: \"kubernetes.io/projected/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-kube-api-access-8pkrd\") pod \"frr-k8s-webhook-server-7fcb986d4-mq57q\" (UID: \"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.276011 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-frr-conf\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.276026 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/968d4f67-d25c-48f4-b120-a345dc551e55-frr-startup\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.276043 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-metrics\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.276063 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mq57q\" (UID: \"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.276080 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-reloader\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.301836 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-4s5lm"] Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.303033 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.305344 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-k5vzm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.313453 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-pb5w7"] Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.314458 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.315220 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.315416 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.315585 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.316096 4677 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.324920 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-pb5w7"] Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377536 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-metrics-certs\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377597 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377652 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xf46\" (UniqueName: \"kubernetes.io/projected/ae71aaf3-8251-4388-ac3a-26121f63a2c0-kube-api-access-2xf46\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377683 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/968d4f67-d25c-48f4-b120-a345dc551e55-metrics-certs\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377708 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6hj\" (UniqueName: \"kubernetes.io/projected/630fc3ca-d708-4a40-a6a0-dc2047d01769-kube-api-access-7c6hj\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377733 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnp5x\" (UniqueName: \"kubernetes.io/projected/968d4f67-d25c-48f4-b120-a345dc551e55-kube-api-access-pnp5x\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377763 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-frr-sockets\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377793 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-cert\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377839 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pkrd\" (UniqueName: \"kubernetes.io/projected/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-kube-api-access-8pkrd\") pod \"frr-k8s-webhook-server-7fcb986d4-mq57q\" (UID: \"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377868 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-frr-conf\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377910 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/968d4f67-d25c-48f4-b120-a345dc551e55-frr-startup\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377942 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-metrics\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.377984 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-metrics-certs\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.378006 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/630fc3ca-d708-4a40-a6a0-dc2047d01769-metallb-excludel2\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.378031 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mq57q\" (UID: \"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.378053 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-reloader\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.378194 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-frr-sockets\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.378303 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-frr-conf\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.378325 4677 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.378374 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-metrics\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.378426 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-cert podName:b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb nodeName:}" failed. No retries permitted until 2025-12-03 14:04:45.878401115 +0000 UTC m=+1076.624733570 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-cert") pod "frr-k8s-webhook-server-7fcb986d4-mq57q" (UID: "b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb") : secret "frr-k8s-webhook-server-cert" not found Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.378431 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/968d4f67-d25c-48f4-b120-a345dc551e55-reloader\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.379431 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/968d4f67-d25c-48f4-b120-a345dc551e55-frr-startup\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.384421 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/968d4f67-d25c-48f4-b120-a345dc551e55-metrics-certs\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.402404 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnp5x\" (UniqueName: \"kubernetes.io/projected/968d4f67-d25c-48f4-b120-a345dc551e55-kube-api-access-pnp5x\") pod \"frr-k8s-lcgmk\" (UID: \"968d4f67-d25c-48f4-b120-a345dc551e55\") " pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.402498 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pkrd\" (UniqueName: \"kubernetes.io/projected/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-kube-api-access-8pkrd\") pod \"frr-k8s-webhook-server-7fcb986d4-mq57q\" (UID: \"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.479511 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.479809 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xf46\" (UniqueName: \"kubernetes.io/projected/ae71aaf3-8251-4388-ac3a-26121f63a2c0-kube-api-access-2xf46\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.479844 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6hj\" (UniqueName: \"kubernetes.io/projected/630fc3ca-d708-4a40-a6a0-dc2047d01769-kube-api-access-7c6hj\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.479875 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-cert\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.479881 4677 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.479961 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist podName:630fc3ca-d708-4a40-a6a0-dc2047d01769 nodeName:}" failed. No retries permitted until 2025-12-03 14:04:45.979924107 +0000 UTC m=+1076.726256562 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist") pod "speaker-4s5lm" (UID: "630fc3ca-d708-4a40-a6a0-dc2047d01769") : secret "metallb-memberlist" not found Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.479983 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-metrics-certs\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.480013 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/630fc3ca-d708-4a40-a6a0-dc2047d01769-metallb-excludel2\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.480059 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-metrics-certs\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.480113 4677 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.480145 4677 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.480173 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-metrics-certs podName:ae71aaf3-8251-4388-ac3a-26121f63a2c0 nodeName:}" failed. No retries permitted until 2025-12-03 14:04:45.980159343 +0000 UTC m=+1076.726491798 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-metrics-certs") pod "controller-f8648f98b-pb5w7" (UID: "ae71aaf3-8251-4388-ac3a-26121f63a2c0") : secret "controller-certs-secret" not found Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.480191 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-metrics-certs podName:630fc3ca-d708-4a40-a6a0-dc2047d01769 nodeName:}" failed. No retries permitted until 2025-12-03 14:04:45.980181973 +0000 UTC m=+1076.726514428 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-metrics-certs") pod "speaker-4s5lm" (UID: "630fc3ca-d708-4a40-a6a0-dc2047d01769") : secret "speaker-certs-secret" not found Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.480756 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/630fc3ca-d708-4a40-a6a0-dc2047d01769-metallb-excludel2\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.483417 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-cert\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.502729 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xf46\" (UniqueName: \"kubernetes.io/projected/ae71aaf3-8251-4388-ac3a-26121f63a2c0-kube-api-access-2xf46\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.505581 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6hj\" (UniqueName: \"kubernetes.io/projected/630fc3ca-d708-4a40-a6a0-dc2047d01769-kube-api-access-7c6hj\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.538791 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.886027 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mq57q\" (UID: \"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.891385 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-mq57q\" (UID: \"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.987830 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.988039 4677 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.988101 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-metrics-certs\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:45 crc kubenswrapper[4677]: E1203 14:04:45.988115 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist podName:630fc3ca-d708-4a40-a6a0-dc2047d01769 nodeName:}" failed. No retries permitted until 2025-12-03 14:04:46.988096362 +0000 UTC m=+1077.734428817 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist") pod "speaker-4s5lm" (UID: "630fc3ca-d708-4a40-a6a0-dc2047d01769") : secret "metallb-memberlist" not found Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.988185 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-metrics-certs\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.992516 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-metrics-certs\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:45 crc kubenswrapper[4677]: I1203 14:04:45.993013 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae71aaf3-8251-4388-ac3a-26121f63a2c0-metrics-certs\") pod \"controller-f8648f98b-pb5w7\" (UID: \"ae71aaf3-8251-4388-ac3a-26121f63a2c0\") " pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:46 crc kubenswrapper[4677]: I1203 14:04:46.100092 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerStarted","Data":"bed5bc4d7d416e8d003ab79883812a80dd636dc40e23814848545cbb480dcada"} Dec 03 14:04:46 crc kubenswrapper[4677]: I1203 14:04:46.147988 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:46 crc kubenswrapper[4677]: I1203 14:04:46.235008 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:46 crc kubenswrapper[4677]: I1203 14:04:46.440119 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q"] Dec 03 14:04:46 crc kubenswrapper[4677]: I1203 14:04:46.538458 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-pb5w7"] Dec 03 14:04:46 crc kubenswrapper[4677]: W1203 14:04:46.546123 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae71aaf3_8251_4388_ac3a_26121f63a2c0.slice/crio-ab6646fa0987c34adab6bfce37026948feac7f4139e76fbf233226780ac5abf8 WatchSource:0}: Error finding container ab6646fa0987c34adab6bfce37026948feac7f4139e76fbf233226780ac5abf8: Status 404 returned error can't find the container with id ab6646fa0987c34adab6bfce37026948feac7f4139e76fbf233226780ac5abf8 Dec 03 14:04:47 crc kubenswrapper[4677]: I1203 14:04:47.007001 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:47 crc kubenswrapper[4677]: I1203 14:04:47.013273 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/630fc3ca-d708-4a40-a6a0-dc2047d01769-memberlist\") pod \"speaker-4s5lm\" (UID: \"630fc3ca-d708-4a40-a6a0-dc2047d01769\") " pod="metallb-system/speaker-4s5lm" Dec 03 14:04:47 crc kubenswrapper[4677]: I1203 14:04:47.105997 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-pb5w7" event={"ID":"ae71aaf3-8251-4388-ac3a-26121f63a2c0","Type":"ContainerStarted","Data":"ab9b36e7bf4712ba1a188df29abc7e2cf65ab5b75664746d4f94be06a72f1e91"} Dec 03 14:04:47 crc kubenswrapper[4677]: I1203 14:04:47.106045 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-pb5w7" event={"ID":"ae71aaf3-8251-4388-ac3a-26121f63a2c0","Type":"ContainerStarted","Data":"ab6646fa0987c34adab6bfce37026948feac7f4139e76fbf233226780ac5abf8"} Dec 03 14:04:47 crc kubenswrapper[4677]: I1203 14:04:47.106787 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" event={"ID":"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb","Type":"ContainerStarted","Data":"651d8fbef1de50024fe61fbad8c3dbad07d8fc5f81916363ccb0bcb2210580be"} Dec 03 14:04:47 crc kubenswrapper[4677]: I1203 14:04:47.125752 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4s5lm" Dec 03 14:04:47 crc kubenswrapper[4677]: W1203 14:04:47.155497 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod630fc3ca_d708_4a40_a6a0_dc2047d01769.slice/crio-45724f4a482e3484023655df3200a9b23665b81e621d0acebacd2898597e76a6 WatchSource:0}: Error finding container 45724f4a482e3484023655df3200a9b23665b81e621d0acebacd2898597e76a6: Status 404 returned error can't find the container with id 45724f4a482e3484023655df3200a9b23665b81e621d0acebacd2898597e76a6 Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.118771 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-pb5w7" event={"ID":"ae71aaf3-8251-4388-ac3a-26121f63a2c0","Type":"ContainerStarted","Data":"b3bcd079b49785930874a2baa290d7699b7423e747d724b562caab37c0f3b522"} Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.119166 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.121022 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4s5lm" event={"ID":"630fc3ca-d708-4a40-a6a0-dc2047d01769","Type":"ContainerStarted","Data":"14212ab3b1b6737a2c76e05bceb0763afaa5c2b9205d96c6f7b4ecc524b2aed2"} Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.121073 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4s5lm" event={"ID":"630fc3ca-d708-4a40-a6a0-dc2047d01769","Type":"ContainerStarted","Data":"c87fa39f636a03990bd57c024b7edd1f8d19bbd43ce1301a4d0817218833c3c5"} Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.121088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4s5lm" event={"ID":"630fc3ca-d708-4a40-a6a0-dc2047d01769","Type":"ContainerStarted","Data":"45724f4a482e3484023655df3200a9b23665b81e621d0acebacd2898597e76a6"} Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.121306 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-4s5lm" Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.144470 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-pb5w7" podStartSLOduration=3.144451598 podStartE2EDuration="3.144451598s" podCreationTimestamp="2025-12-03 14:04:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:04:48.139420489 +0000 UTC m=+1078.885752964" watchObservedRunningTime="2025-12-03 14:04:48.144451598 +0000 UTC m=+1078.890784053" Dec 03 14:04:48 crc kubenswrapper[4677]: I1203 14:04:48.178672 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-4s5lm" podStartSLOduration=3.178650305 podStartE2EDuration="3.178650305s" podCreationTimestamp="2025-12-03 14:04:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:04:48.169416757 +0000 UTC m=+1078.915749212" watchObservedRunningTime="2025-12-03 14:04:48.178650305 +0000 UTC m=+1078.924982760" Dec 03 14:04:56 crc kubenswrapper[4677]: I1203 14:04:56.180340 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" event={"ID":"b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb","Type":"ContainerStarted","Data":"2ff0a3fc9c06ed089f50095841c8ff39ce2d79f713d413b0ac93826ff14d1503"} Dec 03 14:04:56 crc kubenswrapper[4677]: I1203 14:04:56.180978 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:04:56 crc kubenswrapper[4677]: I1203 14:04:56.182146 4677 generic.go:334] "Generic (PLEG): container finished" podID="968d4f67-d25c-48f4-b120-a345dc551e55" containerID="782cf0f4a571ef3a18229b8867bb0051f6730a10a3d7487c22e6ebf49d335fb3" exitCode=0 Dec 03 14:04:56 crc kubenswrapper[4677]: I1203 14:04:56.182190 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerDied","Data":"782cf0f4a571ef3a18229b8867bb0051f6730a10a3d7487c22e6ebf49d335fb3"} Dec 03 14:04:56 crc kubenswrapper[4677]: I1203 14:04:56.240041 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" podStartSLOduration=2.402591944 podStartE2EDuration="11.240019064s" podCreationTimestamp="2025-12-03 14:04:45 +0000 UTC" firstStartedPulling="2025-12-03 14:04:46.45781405 +0000 UTC m=+1077.204146505" lastFinishedPulling="2025-12-03 14:04:55.29524117 +0000 UTC m=+1086.041573625" observedRunningTime="2025-12-03 14:04:56.204797841 +0000 UTC m=+1086.951130296" watchObservedRunningTime="2025-12-03 14:04:56.240019064 +0000 UTC m=+1086.986351559" Dec 03 14:04:56 crc kubenswrapper[4677]: I1203 14:04:56.246255 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-pb5w7" Dec 03 14:04:57 crc kubenswrapper[4677]: I1203 14:04:57.131589 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-4s5lm" Dec 03 14:04:57 crc kubenswrapper[4677]: I1203 14:04:57.195490 4677 generic.go:334] "Generic (PLEG): container finished" podID="968d4f67-d25c-48f4-b120-a345dc551e55" containerID="8dd6901b06ceafe56be451a447ab24162b2d6994e542785c2229bd69f684df3f" exitCode=0 Dec 03 14:04:57 crc kubenswrapper[4677]: I1203 14:04:57.195586 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerDied","Data":"8dd6901b06ceafe56be451a447ab24162b2d6994e542785c2229bd69f684df3f"} Dec 03 14:04:58 crc kubenswrapper[4677]: I1203 14:04:58.203820 4677 generic.go:334] "Generic (PLEG): container finished" podID="968d4f67-d25c-48f4-b120-a345dc551e55" containerID="3976ab015c18c7265876cfdb02e78b9766ff9060b587a887a6e068a05dcbed28" exitCode=0 Dec 03 14:04:58 crc kubenswrapper[4677]: I1203 14:04:58.203924 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerDied","Data":"3976ab015c18c7265876cfdb02e78b9766ff9060b587a887a6e068a05dcbed28"} Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.214190 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerStarted","Data":"99d0488a8cf50e431a5fc8b7d39b1db5a53f8626638e851679a612538a59a3a1"} Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.214500 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.214513 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerStarted","Data":"247e03edf29ddbe611f0a92d239186ad2c293a71929a818940317d2104f14c4c"} Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.214523 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerStarted","Data":"eabe9f171efea1f8ea8e39a49e424725eb18403c51bcc93b2278c7bcc332d196"} Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.214533 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerStarted","Data":"222450654bf760ef6bcfd7c7a4f2739c486052e2bf1465bd04498c2a25d507d6"} Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.214541 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerStarted","Data":"7ec6ce97e06ddfec15976a529fe0186f098b13c08f7a27e701cb0eed329d22d2"} Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.214548 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lcgmk" event={"ID":"968d4f67-d25c-48f4-b120-a345dc551e55","Type":"ContainerStarted","Data":"cf2f9c5f3cde95025aabdb8292d22e5674ddb2223e2fb34bc2e5e06d1df8e012"} Dec 03 14:04:59 crc kubenswrapper[4677]: I1203 14:04:59.238188 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-lcgmk" podStartSLOduration=5.007849177 podStartE2EDuration="14.238173846s" podCreationTimestamp="2025-12-03 14:04:45 +0000 UTC" firstStartedPulling="2025-12-03 14:04:46.073179262 +0000 UTC m=+1076.819511717" lastFinishedPulling="2025-12-03 14:04:55.303503931 +0000 UTC m=+1086.049836386" observedRunningTime="2025-12-03 14:04:59.234821969 +0000 UTC m=+1089.981154444" watchObservedRunningTime="2025-12-03 14:04:59.238173846 +0000 UTC m=+1089.984506301" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.539935 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.578698 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.628056 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6n7ks"] Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.629007 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6n7ks" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.631392 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.636753 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.636830 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-6979x" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.647113 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6n7ks"] Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.697637 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2c2b\" (UniqueName: \"kubernetes.io/projected/f93de765-2a9c-4df6-875f-b346046bed2e-kube-api-access-z2c2b\") pod \"openstack-operator-index-6n7ks\" (UID: \"f93de765-2a9c-4df6-875f-b346046bed2e\") " pod="openstack-operators/openstack-operator-index-6n7ks" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.799036 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2c2b\" (UniqueName: \"kubernetes.io/projected/f93de765-2a9c-4df6-875f-b346046bed2e-kube-api-access-z2c2b\") pod \"openstack-operator-index-6n7ks\" (UID: \"f93de765-2a9c-4df6-875f-b346046bed2e\") " pod="openstack-operators/openstack-operator-index-6n7ks" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.818835 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2c2b\" (UniqueName: \"kubernetes.io/projected/f93de765-2a9c-4df6-875f-b346046bed2e-kube-api-access-z2c2b\") pod \"openstack-operator-index-6n7ks\" (UID: \"f93de765-2a9c-4df6-875f-b346046bed2e\") " pod="openstack-operators/openstack-operator-index-6n7ks" Dec 03 14:05:00 crc kubenswrapper[4677]: I1203 14:05:00.956596 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6n7ks" Dec 03 14:05:01 crc kubenswrapper[4677]: I1203 14:05:01.293027 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6n7ks"] Dec 03 14:05:02 crc kubenswrapper[4677]: I1203 14:05:02.234490 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6n7ks" event={"ID":"f93de765-2a9c-4df6-875f-b346046bed2e","Type":"ContainerStarted","Data":"f8a040be4b39ee2ccbc9c7156c49f7de7f36b60728cea4ae4cb3d22ba995580d"} Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.002850 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6n7ks"] Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.253457 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6n7ks" event={"ID":"f93de765-2a9c-4df6-875f-b346046bed2e","Type":"ContainerStarted","Data":"883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a"} Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.253610 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-6n7ks" podUID="f93de765-2a9c-4df6-875f-b346046bed2e" containerName="registry-server" containerID="cri-o://883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a" gracePeriod=2 Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.277258 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6n7ks" podStartSLOduration=1.63477711 podStartE2EDuration="4.277234664s" podCreationTimestamp="2025-12-03 14:05:00 +0000 UTC" firstStartedPulling="2025-12-03 14:05:01.306567668 +0000 UTC m=+1092.052900123" lastFinishedPulling="2025-12-03 14:05:03.949025222 +0000 UTC m=+1094.695357677" observedRunningTime="2025-12-03 14:05:04.271363614 +0000 UTC m=+1095.017696069" watchObservedRunningTime="2025-12-03 14:05:04.277234664 +0000 UTC m=+1095.023567159" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.609749 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-t5d47"] Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.610921 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.615335 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t5d47"] Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.646584 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2dm9\" (UniqueName: \"kubernetes.io/projected/d21b8d21-356b-4cb9-ab6e-1a08320686fc-kube-api-access-x2dm9\") pod \"openstack-operator-index-t5d47\" (UID: \"d21b8d21-356b-4cb9-ab6e-1a08320686fc\") " pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.675000 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6n7ks" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.757129 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2dm9\" (UniqueName: \"kubernetes.io/projected/d21b8d21-356b-4cb9-ab6e-1a08320686fc-kube-api-access-x2dm9\") pod \"openstack-operator-index-t5d47\" (UID: \"d21b8d21-356b-4cb9-ab6e-1a08320686fc\") " pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.783236 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2dm9\" (UniqueName: \"kubernetes.io/projected/d21b8d21-356b-4cb9-ab6e-1a08320686fc-kube-api-access-x2dm9\") pod \"openstack-operator-index-t5d47\" (UID: \"d21b8d21-356b-4cb9-ab6e-1a08320686fc\") " pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.858007 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2c2b\" (UniqueName: \"kubernetes.io/projected/f93de765-2a9c-4df6-875f-b346046bed2e-kube-api-access-z2c2b\") pod \"f93de765-2a9c-4df6-875f-b346046bed2e\" (UID: \"f93de765-2a9c-4df6-875f-b346046bed2e\") " Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.861006 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f93de765-2a9c-4df6-875f-b346046bed2e-kube-api-access-z2c2b" (OuterVolumeSpecName: "kube-api-access-z2c2b") pod "f93de765-2a9c-4df6-875f-b346046bed2e" (UID: "f93de765-2a9c-4df6-875f-b346046bed2e"). InnerVolumeSpecName "kube-api-access-z2c2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.959641 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2c2b\" (UniqueName: \"kubernetes.io/projected/f93de765-2a9c-4df6-875f-b346046bed2e-kube-api-access-z2c2b\") on node \"crc\" DevicePath \"\"" Dec 03 14:05:04 crc kubenswrapper[4677]: I1203 14:05:04.968551 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.261486 4677 generic.go:334] "Generic (PLEG): container finished" podID="f93de765-2a9c-4df6-875f-b346046bed2e" containerID="883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a" exitCode=0 Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.261537 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6n7ks" event={"ID":"f93de765-2a9c-4df6-875f-b346046bed2e","Type":"ContainerDied","Data":"883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a"} Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.261581 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6n7ks" Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.261613 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6n7ks" event={"ID":"f93de765-2a9c-4df6-875f-b346046bed2e","Type":"ContainerDied","Data":"f8a040be4b39ee2ccbc9c7156c49f7de7f36b60728cea4ae4cb3d22ba995580d"} Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.261641 4677 scope.go:117] "RemoveContainer" containerID="883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a" Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.278271 4677 scope.go:117] "RemoveContainer" containerID="883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a" Dec 03 14:05:05 crc kubenswrapper[4677]: E1203 14:05:05.278620 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a\": container with ID starting with 883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a not found: ID does not exist" containerID="883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a" Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.278655 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a"} err="failed to get container status \"883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a\": rpc error: code = NotFound desc = could not find container \"883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a\": container with ID starting with 883ac9ff1d060b681829a4433b70c79be73850226d2d344ceea9ce4bfee03c4a not found: ID does not exist" Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.304136 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6n7ks"] Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.312000 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-6n7ks"] Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.380423 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t5d47"] Dec 03 14:05:05 crc kubenswrapper[4677]: W1203 14:05:05.386143 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd21b8d21_356b_4cb9_ab6e_1a08320686fc.slice/crio-ce2c52d8458191c7fcb846bc1380e0a4d45fe5be070116422aacbd7c27a9ffb9 WatchSource:0}: Error finding container ce2c52d8458191c7fcb846bc1380e0a4d45fe5be070116422aacbd7c27a9ffb9: Status 404 returned error can't find the container with id ce2c52d8458191c7fcb846bc1380e0a4d45fe5be070116422aacbd7c27a9ffb9 Dec 03 14:05:05 crc kubenswrapper[4677]: I1203 14:05:05.997547 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f93de765-2a9c-4df6-875f-b346046bed2e" path="/var/lib/kubelet/pods/f93de765-2a9c-4df6-875f-b346046bed2e/volumes" Dec 03 14:05:06 crc kubenswrapper[4677]: I1203 14:05:06.153296 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-mq57q" Dec 03 14:05:06 crc kubenswrapper[4677]: I1203 14:05:06.269769 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t5d47" event={"ID":"d21b8d21-356b-4cb9-ab6e-1a08320686fc","Type":"ContainerStarted","Data":"3a7ec5356bfa8bd5a6acdd41a9963102574289ad3ee22bd35e71378505c40ab3"} Dec 03 14:05:06 crc kubenswrapper[4677]: I1203 14:05:06.269815 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t5d47" event={"ID":"d21b8d21-356b-4cb9-ab6e-1a08320686fc","Type":"ContainerStarted","Data":"ce2c52d8458191c7fcb846bc1380e0a4d45fe5be070116422aacbd7c27a9ffb9"} Dec 03 14:05:06 crc kubenswrapper[4677]: I1203 14:05:06.287357 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-t5d47" podStartSLOduration=2.234106189 podStartE2EDuration="2.287337503s" podCreationTimestamp="2025-12-03 14:05:04 +0000 UTC" firstStartedPulling="2025-12-03 14:05:05.389171903 +0000 UTC m=+1096.135504358" lastFinishedPulling="2025-12-03 14:05:05.442403217 +0000 UTC m=+1096.188735672" observedRunningTime="2025-12-03 14:05:06.287282621 +0000 UTC m=+1097.033615076" watchObservedRunningTime="2025-12-03 14:05:06.287337503 +0000 UTC m=+1097.033669978" Dec 03 14:05:14 crc kubenswrapper[4677]: I1203 14:05:14.968847 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:14 crc kubenswrapper[4677]: I1203 14:05:14.970196 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:15 crc kubenswrapper[4677]: I1203 14:05:15.002723 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:15 crc kubenswrapper[4677]: I1203 14:05:15.351690 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-t5d47" Dec 03 14:05:15 crc kubenswrapper[4677]: I1203 14:05:15.542328 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-lcgmk" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.069896 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj"] Dec 03 14:05:16 crc kubenswrapper[4677]: E1203 14:05:16.070133 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93de765-2a9c-4df6-875f-b346046bed2e" containerName="registry-server" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.070146 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93de765-2a9c-4df6-875f-b346046bed2e" containerName="registry-server" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.070285 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93de765-2a9c-4df6-875f-b346046bed2e" containerName="registry-server" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.071118 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.073171 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-vqhqb" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.079263 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj"] Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.109312 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4p87\" (UniqueName: \"kubernetes.io/projected/f962a042-cf75-46fe-8fb1-3c04aaa8043b-kube-api-access-h4p87\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.109530 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-util\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.109719 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-bundle\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.211205 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4p87\" (UniqueName: \"kubernetes.io/projected/f962a042-cf75-46fe-8fb1-3c04aaa8043b-kube-api-access-h4p87\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.211309 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-util\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.211349 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-bundle\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.211771 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-util\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.211852 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-bundle\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.231097 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4p87\" (UniqueName: \"kubernetes.io/projected/f962a042-cf75-46fe-8fb1-3c04aaa8043b-kube-api-access-h4p87\") pod \"6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.386394 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:16 crc kubenswrapper[4677]: I1203 14:05:16.886068 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj"] Dec 03 14:05:16 crc kubenswrapper[4677]: W1203 14:05:16.893340 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf962a042_cf75_46fe_8fb1_3c04aaa8043b.slice/crio-1c14123928be359b818e4e5671f57ea148e9c5f2377896e2cd21914baa08df19 WatchSource:0}: Error finding container 1c14123928be359b818e4e5671f57ea148e9c5f2377896e2cd21914baa08df19: Status 404 returned error can't find the container with id 1c14123928be359b818e4e5671f57ea148e9c5f2377896e2cd21914baa08df19 Dec 03 14:05:17 crc kubenswrapper[4677]: I1203 14:05:17.335088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" event={"ID":"f962a042-cf75-46fe-8fb1-3c04aaa8043b","Type":"ContainerStarted","Data":"1c14123928be359b818e4e5671f57ea148e9c5f2377896e2cd21914baa08df19"} Dec 03 14:05:18 crc kubenswrapper[4677]: I1203 14:05:18.342262 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" event={"ID":"f962a042-cf75-46fe-8fb1-3c04aaa8043b","Type":"ContainerStarted","Data":"50d014e5dcd1493c675164c6c254c4138444626c63d2a149a81752fc8fd2bb5a"} Dec 03 14:05:19 crc kubenswrapper[4677]: I1203 14:05:19.349083 4677 generic.go:334] "Generic (PLEG): container finished" podID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerID="50d014e5dcd1493c675164c6c254c4138444626c63d2a149a81752fc8fd2bb5a" exitCode=0 Dec 03 14:05:19 crc kubenswrapper[4677]: I1203 14:05:19.349140 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" event={"ID":"f962a042-cf75-46fe-8fb1-3c04aaa8043b","Type":"ContainerDied","Data":"50d014e5dcd1493c675164c6c254c4138444626c63d2a149a81752fc8fd2bb5a"} Dec 03 14:05:21 crc kubenswrapper[4677]: I1203 14:05:21.361543 4677 generic.go:334] "Generic (PLEG): container finished" podID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerID="e3632c8f5e36a882c5af9f8c36e52c605d86aa75f56cb9ea033e5e29077ca7e3" exitCode=0 Dec 03 14:05:21 crc kubenswrapper[4677]: I1203 14:05:21.361598 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" event={"ID":"f962a042-cf75-46fe-8fb1-3c04aaa8043b","Type":"ContainerDied","Data":"e3632c8f5e36a882c5af9f8c36e52c605d86aa75f56cb9ea033e5e29077ca7e3"} Dec 03 14:05:22 crc kubenswrapper[4677]: I1203 14:05:22.371079 4677 generic.go:334] "Generic (PLEG): container finished" podID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerID="f63b4aa6b1a6bcdef256c383b741d6acd7a285a4519f11608da16a2316937913" exitCode=0 Dec 03 14:05:22 crc kubenswrapper[4677]: I1203 14:05:22.371185 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" event={"ID":"f962a042-cf75-46fe-8fb1-3c04aaa8043b","Type":"ContainerDied","Data":"f63b4aa6b1a6bcdef256c383b741d6acd7a285a4519f11608da16a2316937913"} Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.037212 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.185297 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4p87\" (UniqueName: \"kubernetes.io/projected/f962a042-cf75-46fe-8fb1-3c04aaa8043b-kube-api-access-h4p87\") pod \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.185376 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-bundle\") pod \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.185488 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-util\") pod \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\" (UID: \"f962a042-cf75-46fe-8fb1-3c04aaa8043b\") " Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.186185 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-bundle" (OuterVolumeSpecName: "bundle") pod "f962a042-cf75-46fe-8fb1-3c04aaa8043b" (UID: "f962a042-cf75-46fe-8fb1-3c04aaa8043b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.195351 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-util" (OuterVolumeSpecName: "util") pod "f962a042-cf75-46fe-8fb1-3c04aaa8043b" (UID: "f962a042-cf75-46fe-8fb1-3c04aaa8043b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.201114 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f962a042-cf75-46fe-8fb1-3c04aaa8043b-kube-api-access-h4p87" (OuterVolumeSpecName: "kube-api-access-h4p87") pod "f962a042-cf75-46fe-8fb1-3c04aaa8043b" (UID: "f962a042-cf75-46fe-8fb1-3c04aaa8043b"). InnerVolumeSpecName "kube-api-access-h4p87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.286470 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4p87\" (UniqueName: \"kubernetes.io/projected/f962a042-cf75-46fe-8fb1-3c04aaa8043b-kube-api-access-h4p87\") on node \"crc\" DevicePath \"\"" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.286505 4677 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.286517 4677 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f962a042-cf75-46fe-8fb1-3c04aaa8043b-util\") on node \"crc\" DevicePath \"\"" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.388312 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" event={"ID":"f962a042-cf75-46fe-8fb1-3c04aaa8043b","Type":"ContainerDied","Data":"1c14123928be359b818e4e5671f57ea148e9c5f2377896e2cd21914baa08df19"} Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.388353 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c14123928be359b818e4e5671f57ea148e9c5f2377896e2cd21914baa08df19" Dec 03 14:05:24 crc kubenswrapper[4677]: I1203 14:05:24.388368 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.154163 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x"] Dec 03 14:05:29 crc kubenswrapper[4677]: E1203 14:05:29.154735 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerName="extract" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.154746 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerName="extract" Dec 03 14:05:29 crc kubenswrapper[4677]: E1203 14:05:29.154760 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerName="util" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.154766 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerName="util" Dec 03 14:05:29 crc kubenswrapper[4677]: E1203 14:05:29.154786 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerName="pull" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.154792 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerName="pull" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.154891 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f962a042-cf75-46fe-8fb1-3c04aaa8043b" containerName="extract" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.155400 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.157524 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-7gvzv" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.195636 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x"] Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.348687 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxq7f\" (UniqueName: \"kubernetes.io/projected/6c72cd7c-f759-4346-a975-4f1cbec9cbf7-kube-api-access-zxq7f\") pod \"openstack-operator-controller-operator-7b44f76b7c-m9c5x\" (UID: \"6c72cd7c-f759-4346-a975-4f1cbec9cbf7\") " pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.449583 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxq7f\" (UniqueName: \"kubernetes.io/projected/6c72cd7c-f759-4346-a975-4f1cbec9cbf7-kube-api-access-zxq7f\") pod \"openstack-operator-controller-operator-7b44f76b7c-m9c5x\" (UID: \"6c72cd7c-f759-4346-a975-4f1cbec9cbf7\") " pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.482516 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxq7f\" (UniqueName: \"kubernetes.io/projected/6c72cd7c-f759-4346-a975-4f1cbec9cbf7-kube-api-access-zxq7f\") pod \"openstack-operator-controller-operator-7b44f76b7c-m9c5x\" (UID: \"6c72cd7c-f759-4346-a975-4f1cbec9cbf7\") " pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:05:29 crc kubenswrapper[4677]: I1203 14:05:29.771970 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:05:30 crc kubenswrapper[4677]: I1203 14:05:30.213892 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x"] Dec 03 14:05:30 crc kubenswrapper[4677]: I1203 14:05:30.223792 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:05:30 crc kubenswrapper[4677]: I1203 14:05:30.453879 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" event={"ID":"6c72cd7c-f759-4346-a975-4f1cbec9cbf7","Type":"ContainerStarted","Data":"18388b55a9da183a001e3d9b8bb589a6690c9696f8d5c4fd81920f5dc101f759"} Dec 03 14:05:34 crc kubenswrapper[4677]: I1203 14:05:34.482616 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" event={"ID":"6c72cd7c-f759-4346-a975-4f1cbec9cbf7","Type":"ContainerStarted","Data":"3b084af4f4ea3e9b7592fedaec83f68517ea892e15d18345595da484a837fa6f"} Dec 03 14:05:34 crc kubenswrapper[4677]: I1203 14:05:34.483243 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:05:34 crc kubenswrapper[4677]: I1203 14:05:34.513454 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" podStartSLOduration=1.434593827 podStartE2EDuration="5.513434306s" podCreationTimestamp="2025-12-03 14:05:29 +0000 UTC" firstStartedPulling="2025-12-03 14:05:30.223447305 +0000 UTC m=+1120.969779760" lastFinishedPulling="2025-12-03 14:05:34.302287784 +0000 UTC m=+1125.048620239" observedRunningTime="2025-12-03 14:05:34.509980257 +0000 UTC m=+1125.256312722" watchObservedRunningTime="2025-12-03 14:05:34.513434306 +0000 UTC m=+1125.259766791" Dec 03 14:05:38 crc kubenswrapper[4677]: I1203 14:05:38.437529 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:05:38 crc kubenswrapper[4677]: I1203 14:05:38.438101 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:05:39 crc kubenswrapper[4677]: I1203 14:05:39.774573 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.832387 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.833731 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.840776 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.841721 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.842340 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-npnll" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.846212 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-hwcgs" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.857585 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.858465 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.861893 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9kzl2" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.871298 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.871610 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.919153 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.920161 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.922296 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-t4w68" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.925022 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.961443 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.975669 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rs4n\" (UniqueName: \"kubernetes.io/projected/460e4110-77d5-476e-88d3-d9ccec539f98-kube-api-access-8rs4n\") pod \"cinder-operator-controller-manager-859b6ccc6-x2q6z\" (UID: \"460e4110-77d5-476e-88d3-d9ccec539f98\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.975741 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j7kj\" (UniqueName: \"kubernetes.io/projected/5236b4d8-b675-4e53-9c7c-c33606436dff-kube-api-access-2j7kj\") pod \"designate-operator-controller-manager-78b4bc895b-45txl\" (UID: \"5236b4d8-b675-4e53-9c7c-c33606436dff\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.976313 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76xcf\" (UniqueName: \"kubernetes.io/projected/df1820a3-7030-449d-ad5d-c8bbdb75ad28-kube-api-access-76xcf\") pod \"barbican-operator-controller-manager-7d9dfd778-jxph7\" (UID: \"df1820a3-7030-449d-ad5d-c8bbdb75ad28\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.976597 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf"] Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.977535 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:06:02 crc kubenswrapper[4677]: I1203 14:06:02.981964 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-lcw62" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.002876 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.077100 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rs4n\" (UniqueName: \"kubernetes.io/projected/460e4110-77d5-476e-88d3-d9ccec539f98-kube-api-access-8rs4n\") pod \"cinder-operator-controller-manager-859b6ccc6-x2q6z\" (UID: \"460e4110-77d5-476e-88d3-d9ccec539f98\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.077191 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbcfh\" (UniqueName: \"kubernetes.io/projected/226bed84-3bd3-4f3b-ad1a-ab8030a71b34-kube-api-access-gbcfh\") pod \"glance-operator-controller-manager-77987cd8cd-8rll8\" (UID: \"226bed84-3bd3-4f3b-ad1a-ab8030a71b34\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.077270 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j7kj\" (UniqueName: \"kubernetes.io/projected/5236b4d8-b675-4e53-9c7c-c33606436dff-kube-api-access-2j7kj\") pod \"designate-operator-controller-manager-78b4bc895b-45txl\" (UID: \"5236b4d8-b675-4e53-9c7c-c33606436dff\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.077319 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76xcf\" (UniqueName: \"kubernetes.io/projected/df1820a3-7030-449d-ad5d-c8bbdb75ad28-kube-api-access-76xcf\") pod \"barbican-operator-controller-manager-7d9dfd778-jxph7\" (UID: \"df1820a3-7030-449d-ad5d-c8bbdb75ad28\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.077416 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26v7x\" (UniqueName: \"kubernetes.io/projected/8d9537f6-e3da-4acb-af76-4ad3bbc403ab-kube-api-access-26v7x\") pod \"heat-operator-controller-manager-5f64f6f8bb-vr9nf\" (UID: \"8d9537f6-e3da-4acb-af76-4ad3bbc403ab\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.110062 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rs4n\" (UniqueName: \"kubernetes.io/projected/460e4110-77d5-476e-88d3-d9ccec539f98-kube-api-access-8rs4n\") pod \"cinder-operator-controller-manager-859b6ccc6-x2q6z\" (UID: \"460e4110-77d5-476e-88d3-d9ccec539f98\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.121502 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j7kj\" (UniqueName: \"kubernetes.io/projected/5236b4d8-b675-4e53-9c7c-c33606436dff-kube-api-access-2j7kj\") pod \"designate-operator-controller-manager-78b4bc895b-45txl\" (UID: \"5236b4d8-b675-4e53-9c7c-c33606436dff\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.125659 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.133796 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76xcf\" (UniqueName: \"kubernetes.io/projected/df1820a3-7030-449d-ad5d-c8bbdb75ad28-kube-api-access-76xcf\") pod \"barbican-operator-controller-manager-7d9dfd778-jxph7\" (UID: \"df1820a3-7030-449d-ad5d-c8bbdb75ad28\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.140310 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.144778 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ssm74" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.158493 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.159806 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.168155 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-5z9r9" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.178760 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26v7x\" (UniqueName: \"kubernetes.io/projected/8d9537f6-e3da-4acb-af76-4ad3bbc403ab-kube-api-access-26v7x\") pod \"heat-operator-controller-manager-5f64f6f8bb-vr9nf\" (UID: \"8d9537f6-e3da-4acb-af76-4ad3bbc403ab\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.178829 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbcfh\" (UniqueName: \"kubernetes.io/projected/226bed84-3bd3-4f3b-ad1a-ab8030a71b34-kube-api-access-gbcfh\") pod \"glance-operator-controller-manager-77987cd8cd-8rll8\" (UID: \"226bed84-3bd3-4f3b-ad1a-ab8030a71b34\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.188290 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.189645 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.191707 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-rdbdw" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.193993 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.195048 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.200647 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-ccnj9" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.221431 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.221426 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbcfh\" (UniqueName: \"kubernetes.io/projected/226bed84-3bd3-4f3b-ad1a-ab8030a71b34-kube-api-access-gbcfh\") pod \"glance-operator-controller-manager-77987cd8cd-8rll8\" (UID: \"226bed84-3bd3-4f3b-ad1a-ab8030a71b34\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.228511 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.236105 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.278453 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.279458 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26v7x\" (UniqueName: \"kubernetes.io/projected/8d9537f6-e3da-4acb-af76-4ad3bbc403ab-kube-api-access-26v7x\") pod \"heat-operator-controller-manager-5f64f6f8bb-vr9nf\" (UID: \"8d9537f6-e3da-4acb-af76-4ad3bbc403ab\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.282767 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.284356 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp4sc\" (UniqueName: \"kubernetes.io/projected/cca481a1-d874-446a-9f63-926247653192-kube-api-access-jp4sc\") pod \"ironic-operator-controller-manager-6c548fd776-rcgbn\" (UID: \"cca481a1-d874-446a-9f63-926247653192\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.284450 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnrj2\" (UniqueName: \"kubernetes.io/projected/829c01d5-d4d0-40a4-9513-0e60b322d4d0-kube-api-access-rnrj2\") pod \"manila-operator-controller-manager-7c79b5df47-l2wcn\" (UID: \"829c01d5-d4d0-40a4-9513-0e60b322d4d0\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.284643 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5clr\" (UniqueName: \"kubernetes.io/projected/af346d56-cdf6-408d-87fc-6431b12a9cd8-kube-api-access-n5clr\") pod \"keystone-operator-controller-manager-7765d96ddf-5f5nf\" (UID: \"af346d56-cdf6-408d-87fc-6431b12a9cd8\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.284675 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mp82\" (UniqueName: \"kubernetes.io/projected/0ffc0c32-9591-4a0d-b7fb-10b8faa85c18-kube-api-access-5mp82\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-n6xjd\" (UID: \"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.284726 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqtwk\" (UniqueName: \"kubernetes.io/projected/01e63cef-045a-42e0-8776-2f07b1187a40-kube-api-access-zqtwk\") pod \"horizon-operator-controller-manager-68c6d99b8f-kqdd2\" (UID: \"01e63cef-045a-42e0-8776-2f07b1187a40\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.284808 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.293690 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-w484t" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.304105 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.305746 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.310363 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.313487 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-j92qt" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.352499 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.353868 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.357440 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vnwh6" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.359407 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.361560 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.377129 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.391443 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnrj2\" (UniqueName: \"kubernetes.io/projected/829c01d5-d4d0-40a4-9513-0e60b322d4d0-kube-api-access-rnrj2\") pod \"manila-operator-controller-manager-7c79b5df47-l2wcn\" (UID: \"829c01d5-d4d0-40a4-9513-0e60b322d4d0\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.391675 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5clr\" (UniqueName: \"kubernetes.io/projected/af346d56-cdf6-408d-87fc-6431b12a9cd8-kube-api-access-n5clr\") pod \"keystone-operator-controller-manager-7765d96ddf-5f5nf\" (UID: \"af346d56-cdf6-408d-87fc-6431b12a9cd8\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.391700 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mp82\" (UniqueName: \"kubernetes.io/projected/0ffc0c32-9591-4a0d-b7fb-10b8faa85c18-kube-api-access-5mp82\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-n6xjd\" (UID: \"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.391739 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqtwk\" (UniqueName: \"kubernetes.io/projected/01e63cef-045a-42e0-8776-2f07b1187a40-kube-api-access-zqtwk\") pod \"horizon-operator-controller-manager-68c6d99b8f-kqdd2\" (UID: \"01e63cef-045a-42e0-8776-2f07b1187a40\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.391768 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp4sc\" (UniqueName: \"kubernetes.io/projected/cca481a1-d874-446a-9f63-926247653192-kube-api-access-jp4sc\") pod \"ironic-operator-controller-manager-6c548fd776-rcgbn\" (UID: \"cca481a1-d874-446a-9f63-926247653192\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.401589 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.417076 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.427617 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.428712 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5clr\" (UniqueName: \"kubernetes.io/projected/af346d56-cdf6-408d-87fc-6431b12a9cd8-kube-api-access-n5clr\") pod \"keystone-operator-controller-manager-7765d96ddf-5f5nf\" (UID: \"af346d56-cdf6-408d-87fc-6431b12a9cd8\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.434590 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnrj2\" (UniqueName: \"kubernetes.io/projected/829c01d5-d4d0-40a4-9513-0e60b322d4d0-kube-api-access-rnrj2\") pod \"manila-operator-controller-manager-7c79b5df47-l2wcn\" (UID: \"829c01d5-d4d0-40a4-9513-0e60b322d4d0\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.434842 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.435126 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp4sc\" (UniqueName: \"kubernetes.io/projected/cca481a1-d874-446a-9f63-926247653192-kube-api-access-jp4sc\") pod \"ironic-operator-controller-manager-6c548fd776-rcgbn\" (UID: \"cca481a1-d874-446a-9f63-926247653192\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.442305 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mp82\" (UniqueName: \"kubernetes.io/projected/0ffc0c32-9591-4a0d-b7fb-10b8faa85c18-kube-api-access-5mp82\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-n6xjd\" (UID: \"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.444703 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.444874 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.445326 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqtwk\" (UniqueName: \"kubernetes.io/projected/01e63cef-045a-42e0-8776-2f07b1187a40-kube-api-access-zqtwk\") pod \"horizon-operator-controller-manager-68c6d99b8f-kqdd2\" (UID: \"01e63cef-045a-42e0-8776-2f07b1187a40\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.448465 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.449516 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-wnp9v" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.461832 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.463019 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.466003 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-xt27n" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.467525 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.472397 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.476068 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.476176 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.486147 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.486717 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-wfv7x" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.487175 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.488296 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.490116 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-27d4p" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.498455 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fjq9\" (UniqueName: \"kubernetes.io/projected/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-kube-api-access-2fjq9\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.498567 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.498602 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgpzj\" (UniqueName: \"kubernetes.io/projected/1e0b7fb1-cdb4-44a9-9508-4939f1038023-kube-api-access-qgpzj\") pod \"mariadb-operator-controller-manager-56bbcc9d85-4tw2x\" (UID: \"1e0b7fb1-cdb4-44a9-9508-4939f1038023\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.498750 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.499925 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.501023 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.501100 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.503489 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-8gr7d" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.504082 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.511195 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.515649 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.516648 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.519514 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-6g62z" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.527163 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.537471 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.540144 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.544053 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.545302 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.556462 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-949ww" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.556673 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-crsk5" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.560019 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.570869 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.595439 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.597168 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601091 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601177 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgpzj\" (UniqueName: \"kubernetes.io/projected/1e0b7fb1-cdb4-44a9-9508-4939f1038023-kube-api-access-qgpzj\") pod \"mariadb-operator-controller-manager-56bbcc9d85-4tw2x\" (UID: \"1e0b7fb1-cdb4-44a9-9508-4939f1038023\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601238 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnrt8\" (UniqueName: \"kubernetes.io/projected/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-kube-api-access-lnrt8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601289 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601331 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdxhd\" (UniqueName: \"kubernetes.io/projected/da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6-kube-api-access-tdxhd\") pod \"ovn-operator-controller-manager-b6456fdb6-hfwgd\" (UID: \"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601422 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxs2n\" (UniqueName: \"kubernetes.io/projected/59757e80-419c-40f9-9ab2-f6dbeff58f7c-kube-api-access-lxs2n\") pod \"octavia-operator-controller-manager-998648c74-vv4wp\" (UID: \"59757e80-419c-40f9-9ab2-f6dbeff58f7c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601474 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fjq9\" (UniqueName: \"kubernetes.io/projected/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-kube-api-access-2fjq9\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601499 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m97j7\" (UniqueName: \"kubernetes.io/projected/7270260d-7534-4b53-be80-69fbee93fdb7-kube-api-access-m97j7\") pod \"nova-operator-controller-manager-697bc559fc-vk8s9\" (UID: \"7270260d-7534-4b53-be80-69fbee93fdb7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.601645 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:06:03 crc kubenswrapper[4677]: E1203 14:06:03.601741 4677 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:03 crc kubenswrapper[4677]: E1203 14:06:03.601811 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert podName:5b6f5612-8a33-46f2-8f4e-415d5bc2e807 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:04.101790668 +0000 UTC m=+1154.848123123 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert") pod "infra-operator-controller-manager-57548d458d-qhdx7" (UID: "5b6f5612-8a33-46f2-8f4e-415d5bc2e807") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.602893 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ww7ws" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.603160 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.648680 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgpzj\" (UniqueName: \"kubernetes.io/projected/1e0b7fb1-cdb4-44a9-9508-4939f1038023-kube-api-access-qgpzj\") pod \"mariadb-operator-controller-manager-56bbcc9d85-4tw2x\" (UID: \"1e0b7fb1-cdb4-44a9-9508-4939f1038023\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.663024 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.681275 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702450 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnc2c\" (UniqueName: \"kubernetes.io/projected/bbb9ba89-c82a-4cb6-8346-a76d2f24fd38-kube-api-access-qnc2c\") pod \"swift-operator-controller-manager-5f8c65bbfc-8jjpl\" (UID: \"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702516 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7rpj\" (UniqueName: \"kubernetes.io/projected/882195ac-f224-4563-b00d-0291e40f3204-kube-api-access-g7rpj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-9jc84\" (UID: \"882195ac-f224-4563-b00d-0291e40f3204\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702535 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkrcm\" (UniqueName: \"kubernetes.io/projected/47d9316a-afea-441d-8c42-45e90efe4d5a-kube-api-access-zkrcm\") pod \"placement-operator-controller-manager-78f8948974-qxf8w\" (UID: \"47d9316a-afea-441d-8c42-45e90efe4d5a\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702587 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnrt8\" (UniqueName: \"kubernetes.io/projected/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-kube-api-access-lnrt8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702606 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702636 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bjr2\" (UniqueName: \"kubernetes.io/projected/85cba130-9ba2-48b7-917f-14946bc72419-kube-api-access-4bjr2\") pod \"test-operator-controller-manager-5854674fcc-jmxdr\" (UID: \"85cba130-9ba2-48b7-917f-14946bc72419\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702661 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdxhd\" (UniqueName: \"kubernetes.io/projected/da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6-kube-api-access-tdxhd\") pod \"ovn-operator-controller-manager-b6456fdb6-hfwgd\" (UID: \"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702695 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxs2n\" (UniqueName: \"kubernetes.io/projected/59757e80-419c-40f9-9ab2-f6dbeff58f7c-kube-api-access-lxs2n\") pod \"octavia-operator-controller-manager-998648c74-vv4wp\" (UID: \"59757e80-419c-40f9-9ab2-f6dbeff58f7c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702717 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m97j7\" (UniqueName: \"kubernetes.io/projected/7270260d-7534-4b53-be80-69fbee93fdb7-kube-api-access-m97j7\") pod \"nova-operator-controller-manager-697bc559fc-vk8s9\" (UID: \"7270260d-7534-4b53-be80-69fbee93fdb7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.702753 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t9mn\" (UniqueName: \"kubernetes.io/projected/40fcffeb-3d82-431f-a208-c06bc3f30557-kube-api-access-8t9mn\") pod \"watcher-operator-controller-manager-6757bc7c76-rqljj\" (UID: \"40fcffeb-3d82-431f-a208-c06bc3f30557\") " pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:06:03 crc kubenswrapper[4677]: E1203 14:06:03.703649 4677 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:03 crc kubenswrapper[4677]: E1203 14:06:03.704038 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert podName:de88ce12-43f1-4e18-ad73-f7d2c222c4ca nodeName:}" failed. No retries permitted until 2025-12-03 14:06:04.20370132 +0000 UTC m=+1154.950033825 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" (UID: "de88ce12-43f1-4e18-ad73-f7d2c222c4ca") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.710668 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.717297 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fjq9\" (UniqueName: \"kubernetes.io/projected/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-kube-api-access-2fjq9\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.727082 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdxhd\" (UniqueName: \"kubernetes.io/projected/da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6-kube-api-access-tdxhd\") pod \"ovn-operator-controller-manager-b6456fdb6-hfwgd\" (UID: \"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.729191 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m97j7\" (UniqueName: \"kubernetes.io/projected/7270260d-7534-4b53-be80-69fbee93fdb7-kube-api-access-m97j7\") pod \"nova-operator-controller-manager-697bc559fc-vk8s9\" (UID: \"7270260d-7534-4b53-be80-69fbee93fdb7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.729883 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnrt8\" (UniqueName: \"kubernetes.io/projected/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-kube-api-access-lnrt8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.737629 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.754355 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxs2n\" (UniqueName: \"kubernetes.io/projected/59757e80-419c-40f9-9ab2-f6dbeff58f7c-kube-api-access-lxs2n\") pod \"octavia-operator-controller-manager-998648c74-vv4wp\" (UID: \"59757e80-419c-40f9-9ab2-f6dbeff58f7c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.765772 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.780542 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.781740 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.785991 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.786052 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.787115 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5vhvc" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.788270 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.788593 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.804647 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnc2c\" (UniqueName: \"kubernetes.io/projected/bbb9ba89-c82a-4cb6-8346-a76d2f24fd38-kube-api-access-qnc2c\") pod \"swift-operator-controller-manager-5f8c65bbfc-8jjpl\" (UID: \"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.804706 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7rpj\" (UniqueName: \"kubernetes.io/projected/882195ac-f224-4563-b00d-0291e40f3204-kube-api-access-g7rpj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-9jc84\" (UID: \"882195ac-f224-4563-b00d-0291e40f3204\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.804726 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkrcm\" (UniqueName: \"kubernetes.io/projected/47d9316a-afea-441d-8c42-45e90efe4d5a-kube-api-access-zkrcm\") pod \"placement-operator-controller-manager-78f8948974-qxf8w\" (UID: \"47d9316a-afea-441d-8c42-45e90efe4d5a\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.804791 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bjr2\" (UniqueName: \"kubernetes.io/projected/85cba130-9ba2-48b7-917f-14946bc72419-kube-api-access-4bjr2\") pod \"test-operator-controller-manager-5854674fcc-jmxdr\" (UID: \"85cba130-9ba2-48b7-917f-14946bc72419\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.804835 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t9mn\" (UniqueName: \"kubernetes.io/projected/40fcffeb-3d82-431f-a208-c06bc3f30557-kube-api-access-8t9mn\") pod \"watcher-operator-controller-manager-6757bc7c76-rqljj\" (UID: \"40fcffeb-3d82-431f-a208-c06bc3f30557\") " pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.830563 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.831203 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.832224 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.835584 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t9mn\" (UniqueName: \"kubernetes.io/projected/40fcffeb-3d82-431f-a208-c06bc3f30557-kube-api-access-8t9mn\") pod \"watcher-operator-controller-manager-6757bc7c76-rqljj\" (UID: \"40fcffeb-3d82-431f-a208-c06bc3f30557\") " pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.835752 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-ttll5" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.835901 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnc2c\" (UniqueName: \"kubernetes.io/projected/bbb9ba89-c82a-4cb6-8346-a76d2f24fd38-kube-api-access-qnc2c\") pod \"swift-operator-controller-manager-5f8c65bbfc-8jjpl\" (UID: \"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.837258 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7rpj\" (UniqueName: \"kubernetes.io/projected/882195ac-f224-4563-b00d-0291e40f3204-kube-api-access-g7rpj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-9jc84\" (UID: \"882195ac-f224-4563-b00d-0291e40f3204\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.839820 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7"] Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.848355 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkrcm\" (UniqueName: \"kubernetes.io/projected/47d9316a-afea-441d-8c42-45e90efe4d5a-kube-api-access-zkrcm\") pod \"placement-operator-controller-manager-78f8948974-qxf8w\" (UID: \"47d9316a-afea-441d-8c42-45e90efe4d5a\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.848517 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bjr2\" (UniqueName: \"kubernetes.io/projected/85cba130-9ba2-48b7-917f-14946bc72419-kube-api-access-4bjr2\") pod \"test-operator-controller-manager-5854674fcc-jmxdr\" (UID: \"85cba130-9ba2-48b7-917f-14946bc72419\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.854451 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.905600 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.905654 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4gb8\" (UniqueName: \"kubernetes.io/projected/f5d376fc-835d-44d7-954a-f838a22f7aa8-kube-api-access-t4gb8\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.905725 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:03 crc kubenswrapper[4677]: I1203 14:06:03.937150 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.007328 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqfwd\" (UniqueName: \"kubernetes.io/projected/27ee4b91-853f-444b-8663-d913a1d9d1e5-kube-api-access-hqfwd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qnqg7\" (UID: \"27ee4b91-853f-444b-8663-d913a1d9d1e5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.007555 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.007666 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4gb8\" (UniqueName: \"kubernetes.io/projected/f5d376fc-835d-44d7-954a-f838a22f7aa8-kube-api-access-t4gb8\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.007784 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.007995 4677 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.008122 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:04.508105202 +0000 UTC m=+1155.254437657 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.008261 4677 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.008365 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:04.508354518 +0000 UTC m=+1155.254686973 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "metrics-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.031798 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4gb8\" (UniqueName: \"kubernetes.io/projected/f5d376fc-835d-44d7-954a-f838a22f7aa8-kube-api-access-t4gb8\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.035277 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.055303 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.086789 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.125848 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqfwd\" (UniqueName: \"kubernetes.io/projected/27ee4b91-853f-444b-8663-d913a1d9d1e5-kube-api-access-hqfwd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qnqg7\" (UID: \"27ee4b91-853f-444b-8663-d913a1d9d1e5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.125938 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.128181 4677 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.128228 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert podName:5b6f5612-8a33-46f2-8f4e-415d5bc2e807 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:05.128214 +0000 UTC m=+1155.874546455 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert") pod "infra-operator-controller-manager-57548d458d-qhdx7" (UID: "5b6f5612-8a33-46f2-8f4e-415d5bc2e807") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.163060 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqfwd\" (UniqueName: \"kubernetes.io/projected/27ee4b91-853f-444b-8663-d913a1d9d1e5-kube-api-access-hqfwd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qnqg7\" (UID: \"27ee4b91-853f-444b-8663-d913a1d9d1e5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.212806 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.227166 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.227337 4677 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.227392 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert podName:de88ce12-43f1-4e18-ad73-f7d2c222c4ca nodeName:}" failed. No retries permitted until 2025-12-03 14:06:05.227373132 +0000 UTC m=+1155.973705597 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" (UID: "de88ce12-43f1-4e18-ad73-f7d2c222c4ca") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.536294 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7"] Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.549108 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8"] Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.595421 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.595506 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.595642 4677 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.595660 4677 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.595686 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:05.595672001 +0000 UTC m=+1156.342004456 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "metrics-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: E1203 14:06:04.595726 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:05.595706062 +0000 UTC m=+1156.342038567 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "webhook-server-cert" not found Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.654531 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl"] Dec 03 14:06:04 crc kubenswrapper[4677]: W1203 14:06:04.655817 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5236b4d8_b675_4e53_9c7c_c33606436dff.slice/crio-0de5c42e4e04bf7d02a37d862f9e6aacfa9d3e08db060af1639649ee1fed1ff4 WatchSource:0}: Error finding container 0de5c42e4e04bf7d02a37d862f9e6aacfa9d3e08db060af1639649ee1fed1ff4: Status 404 returned error can't find the container with id 0de5c42e4e04bf7d02a37d862f9e6aacfa9d3e08db060af1639649ee1fed1ff4 Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.678180 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn"] Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.744554 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf"] Dec 03 14:06:04 crc kubenswrapper[4677]: W1203 14:06:04.745344 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d9537f6_e3da_4acb_af76_4ad3bbc403ab.slice/crio-71c6785a798d404a979da8cbc233daad3dc9f864a57fab98af2ce62c19f0865f WatchSource:0}: Error finding container 71c6785a798d404a979da8cbc233daad3dc9f864a57fab98af2ce62c19f0865f: Status 404 returned error can't find the container with id 71c6785a798d404a979da8cbc233daad3dc9f864a57fab98af2ce62c19f0865f Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.751223 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" event={"ID":"df1820a3-7030-449d-ad5d-c8bbdb75ad28","Type":"ContainerStarted","Data":"4b719dc4eac6bbeb4bc4072e6315e5b859cb7b8a70280ce05dd117583005afb4"} Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.766706 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" event={"ID":"226bed84-3bd3-4f3b-ad1a-ab8030a71b34","Type":"ContainerStarted","Data":"149972a59e450413ddf6ff52fece74243061e6d0f6a7a92297b04cf0e3a6b3fd"} Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.770533 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2"] Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.770572 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" event={"ID":"5236b4d8-b675-4e53-9c7c-c33606436dff","Type":"ContainerStarted","Data":"0de5c42e4e04bf7d02a37d862f9e6aacfa9d3e08db060af1639649ee1fed1ff4"} Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.772160 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" event={"ID":"cca481a1-d874-446a-9f63-926247653192","Type":"ContainerStarted","Data":"2ac11f6087d25299fe046c4f01abbdb633e46ba91107df4374ac3d25b993ff49"} Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.869013 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd"] Dec 03 14:06:04 crc kubenswrapper[4677]: W1203 14:06:04.870753 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ffc0c32_9591_4a0d_b7fb_10b8faa85c18.slice/crio-d8f29f4d592328bed282300e31269832ef3e4be0fd2d71b9fe2a350085463acf WatchSource:0}: Error finding container d8f29f4d592328bed282300e31269832ef3e4be0fd2d71b9fe2a350085463acf: Status 404 returned error can't find the container with id d8f29f4d592328bed282300e31269832ef3e4be0fd2d71b9fe2a350085463acf Dec 03 14:06:04 crc kubenswrapper[4677]: W1203 14:06:04.872364 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7270260d_7534_4b53_be80_69fbee93fdb7.slice/crio-72e8adef0394ec056f139b0e4ea9368a1d4badb55a0130039e1c4073d129b939 WatchSource:0}: Error finding container 72e8adef0394ec056f139b0e4ea9368a1d4badb55a0130039e1c4073d129b939: Status 404 returned error can't find the container with id 72e8adef0394ec056f139b0e4ea9368a1d4badb55a0130039e1c4073d129b939 Dec 03 14:06:04 crc kubenswrapper[4677]: W1203 14:06:04.878278 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod460e4110_77d5_476e_88d3_d9ccec539f98.slice/crio-0642ad930aeb65388dbf04386be724a01775b9f09033004469ac9c9c9ec9d96b WatchSource:0}: Error finding container 0642ad930aeb65388dbf04386be724a01775b9f09033004469ac9c9c9ec9d96b: Status 404 returned error can't find the container with id 0642ad930aeb65388dbf04386be724a01775b9f09033004469ac9c9c9ec9d96b Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.883709 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9"] Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.889062 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn"] Dec 03 14:06:04 crc kubenswrapper[4677]: I1203 14:06:04.894235 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z"] Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.038811 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x"] Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.058794 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp"] Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.067477 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd"] Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.068158 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tdxhd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-hfwgd_openstack-operators(da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.072734 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf"] Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.086883 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tdxhd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-hfwgd_openstack-operators(da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.089012 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podUID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.094574 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n5clr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-5f5nf_openstack-operators(af346d56-cdf6-408d-87fc-6431b12a9cd8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.096055 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7"] Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.103220 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n5clr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-5f5nf_openstack-operators(af346d56-cdf6-408d-87fc-6431b12a9cd8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.104648 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.105749 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr"] Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.114534 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w"] Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.115564 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4bjr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-jmxdr_openstack-operators(85cba130-9ba2-48b7-917f-14946bc72419): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.117465 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qnc2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-8jjpl_openstack-operators(bbb9ba89-c82a-4cb6-8346-a76d2f24fd38): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.117633 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4bjr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-jmxdr_openstack-operators(85cba130-9ba2-48b7-917f-14946bc72419): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.117680 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zkrcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-qxf8w_openstack-operators(47d9316a-afea-441d-8c42-45e90efe4d5a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.117747 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hqfwd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qnqg7_openstack-operators(27ee4b91-853f-444b-8663-d913a1d9d1e5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.118879 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" podUID="27ee4b91-853f-444b-8663-d913a1d9d1e5" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.119102 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.119192 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84"] Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.121505 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zkrcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-qxf8w_openstack-operators(47d9316a-afea-441d-8c42-45e90efe4d5a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.121798 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qnc2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-8jjpl_openstack-operators(bbb9ba89-c82a-4cb6-8346-a76d2f24fd38): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: W1203 14:06:05.121892 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod882195ac_f224_4563_b00d_0291e40f3204.slice/crio-3292c9c021de34aa0780959a510f3e0310df1271c22faaf14ec4897a5596056e WatchSource:0}: Error finding container 3292c9c021de34aa0780959a510f3e0310df1271c22faaf14ec4897a5596056e: Status 404 returned error can't find the container with id 3292c9c021de34aa0780959a510f3e0310df1271c22faaf14ec4897a5596056e Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.122825 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" podUID="47d9316a-afea-441d-8c42-45e90efe4d5a" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.122901 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.123168 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g7rpj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-9jc84_openstack-operators(882195ac-f224-4563-b00d-0291e40f3204): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.124936 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g7rpj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-9jc84_openstack-operators(882195ac-f224-4563-b00d-0291e40f3204): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.126066 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl"] Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.126142 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podUID="882195ac-f224-4563-b00d-0291e40f3204" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.200420 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj"] Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.207538 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.207698 4677 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.207781 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert podName:5b6f5612-8a33-46f2-8f4e-415d5bc2e807 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:07.207765379 +0000 UTC m=+1157.954097824 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert") pod "infra-operator-controller-manager-57548d458d-qhdx7" (UID: "5b6f5612-8a33-46f2-8f4e-415d5bc2e807") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.309398 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.309715 4677 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.309800 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert podName:de88ce12-43f1-4e18-ad73-f7d2c222c4ca nodeName:}" failed. No retries permitted until 2025-12-03 14:06:07.309777653 +0000 UTC m=+1158.056110168 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" (UID: "de88ce12-43f1-4e18-ad73-f7d2c222c4ca") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.613615 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.613735 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.614336 4677 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.614391 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:07.6143739 +0000 UTC m=+1158.360706365 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "metrics-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.614938 4677 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.614988 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:07.614976635 +0000 UTC m=+1158.361309090 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "webhook-server-cert" not found Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.787296 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" event={"ID":"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18","Type":"ContainerStarted","Data":"d8f29f4d592328bed282300e31269832ef3e4be0fd2d71b9fe2a350085463acf"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.789986 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" event={"ID":"40fcffeb-3d82-431f-a208-c06bc3f30557","Type":"ContainerStarted","Data":"88f6f90690932ec350469f015ee8dd85ea8ece881250444c8f114a11d77ed63c"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.791061 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" event={"ID":"1e0b7fb1-cdb4-44a9-9508-4939f1038023","Type":"ContainerStarted","Data":"b232ae9564d0f388cb394f62924bdabb99bf880c2bec0ba084721a93016b36f7"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.792489 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" event={"ID":"882195ac-f224-4563-b00d-0291e40f3204","Type":"ContainerStarted","Data":"3292c9c021de34aa0780959a510f3e0310df1271c22faaf14ec4897a5596056e"} Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.793724 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podUID="882195ac-f224-4563-b00d-0291e40f3204" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.795072 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" event={"ID":"01e63cef-045a-42e0-8776-2f07b1187a40","Type":"ContainerStarted","Data":"b9cdea93d1e5f0af7d0efa4869667161505f097916e1b39c71686eb47af6cd58"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.799895 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" event={"ID":"460e4110-77d5-476e-88d3-d9ccec539f98","Type":"ContainerStarted","Data":"0642ad930aeb65388dbf04386be724a01775b9f09033004469ac9c9c9ec9d96b"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.801852 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" event={"ID":"85cba130-9ba2-48b7-917f-14946bc72419","Type":"ContainerStarted","Data":"993aea8e621db06a394e78e50b76a0b70566b2f1620fe8785c52cb55cb68b99f"} Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.804633 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.805199 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" event={"ID":"27ee4b91-853f-444b-8663-d913a1d9d1e5","Type":"ContainerStarted","Data":"4ebc3f99df838651bee7282921baaa4ad16af840e0ad945ff058c67d50bb0fa9"} Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.808818 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" podUID="27ee4b91-853f-444b-8663-d913a1d9d1e5" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.810232 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" event={"ID":"829c01d5-d4d0-40a4-9513-0e60b322d4d0","Type":"ContainerStarted","Data":"d496304d355775a24d24917ab827cfb44f2a5d3b272940198b8ec51501b6803d"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.816197 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" event={"ID":"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38","Type":"ContainerStarted","Data":"b69900df15c4193a0a7b240beaa18935b67b99c5c46ca389d47d49675f494483"} Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.822065 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.849198 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" event={"ID":"af346d56-cdf6-408d-87fc-6431b12a9cd8","Type":"ContainerStarted","Data":"c01d53a1729a3a8fc99d1d8ced1239b20ef1ce5b6f3eeed462ac4fc37c96763f"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.853357 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" event={"ID":"8d9537f6-e3da-4acb-af76-4ad3bbc403ab","Type":"ContainerStarted","Data":"71c6785a798d404a979da8cbc233daad3dc9f864a57fab98af2ce62c19f0865f"} Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.853936 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.855835 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" event={"ID":"59757e80-419c-40f9-9ab2-f6dbeff58f7c","Type":"ContainerStarted","Data":"a733e6ec3321a4f39229a0530a8908db388e9bc0d2a60ddd5bc7c6c36eab0417"} Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.865913 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" event={"ID":"47d9316a-afea-441d-8c42-45e90efe4d5a","Type":"ContainerStarted","Data":"e58b6f983f3c5cd4f879fa18a22e1b33b6351dcb5d1638fef30b14e8c678bed7"} Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.869097 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" podUID="47d9316a-afea-441d-8c42-45e90efe4d5a" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.870829 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" event={"ID":"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6","Type":"ContainerStarted","Data":"69f13d52738caf3f17d59b83309e5fae1ab435aa2f4a7cb7a9ebf2d3135dc389"} Dec 03 14:06:05 crc kubenswrapper[4677]: E1203 14:06:05.882201 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podUID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" Dec 03 14:06:05 crc kubenswrapper[4677]: I1203 14:06:05.884765 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" event={"ID":"7270260d-7534-4b53-be80-69fbee93fdb7","Type":"ContainerStarted","Data":"72e8adef0394ec056f139b0e4ea9368a1d4badb55a0130039e1c4073d129b939"} Dec 03 14:06:06 crc kubenswrapper[4677]: E1203 14:06:06.897832 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" Dec 03 14:06:06 crc kubenswrapper[4677]: E1203 14:06:06.898798 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" Dec 03 14:06:06 crc kubenswrapper[4677]: E1203 14:06:06.899050 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" podUID="27ee4b91-853f-444b-8663-d913a1d9d1e5" Dec 03 14:06:06 crc kubenswrapper[4677]: E1203 14:06:06.899348 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" podUID="47d9316a-afea-441d-8c42-45e90efe4d5a" Dec 03 14:06:06 crc kubenswrapper[4677]: E1203 14:06:06.899333 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podUID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" Dec 03 14:06:06 crc kubenswrapper[4677]: E1203 14:06:06.899418 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" Dec 03 14:06:06 crc kubenswrapper[4677]: E1203 14:06:06.899439 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podUID="882195ac-f224-4563-b00d-0291e40f3204" Dec 03 14:06:07 crc kubenswrapper[4677]: I1203 14:06:07.238877 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.239081 4677 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.239312 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert podName:5b6f5612-8a33-46f2-8f4e-415d5bc2e807 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:11.239295346 +0000 UTC m=+1161.985627791 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert") pod "infra-operator-controller-manager-57548d458d-qhdx7" (UID: "5b6f5612-8a33-46f2-8f4e-415d5bc2e807") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:07 crc kubenswrapper[4677]: I1203 14:06:07.341228 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.341410 4677 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.341472 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert podName:de88ce12-43f1-4e18-ad73-f7d2c222c4ca nodeName:}" failed. No retries permitted until 2025-12-03 14:06:11.341457694 +0000 UTC m=+1162.087790149 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" (UID: "de88ce12-43f1-4e18-ad73-f7d2c222c4ca") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:07 crc kubenswrapper[4677]: I1203 14:06:07.646293 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:07 crc kubenswrapper[4677]: I1203 14:06:07.646399 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.646431 4677 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.646497 4677 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.646524 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:11.646506282 +0000 UTC m=+1162.392838737 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "webhook-server-cert" not found Dec 03 14:06:07 crc kubenswrapper[4677]: E1203 14:06:07.646564 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:11.646556074 +0000 UTC m=+1162.392888529 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "metrics-server-cert" not found Dec 03 14:06:08 crc kubenswrapper[4677]: I1203 14:06:08.436839 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:06:08 crc kubenswrapper[4677]: I1203 14:06:08.436888 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:06:11 crc kubenswrapper[4677]: I1203 14:06:11.334145 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.334381 4677 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.334494 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert podName:5b6f5612-8a33-46f2-8f4e-415d5bc2e807 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:19.334478204 +0000 UTC m=+1170.080810659 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert") pod "infra-operator-controller-manager-57548d458d-qhdx7" (UID: "5b6f5612-8a33-46f2-8f4e-415d5bc2e807") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:06:11 crc kubenswrapper[4677]: I1203 14:06:11.436178 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.436394 4677 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.436460 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert podName:de88ce12-43f1-4e18-ad73-f7d2c222c4ca nodeName:}" failed. No retries permitted until 2025-12-03 14:06:19.436443447 +0000 UTC m=+1170.182775902 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" (UID: "de88ce12-43f1-4e18-ad73-f7d2c222c4ca") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:06:11 crc kubenswrapper[4677]: I1203 14:06:11.740792 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:11 crc kubenswrapper[4677]: I1203 14:06:11.740889 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.741023 4677 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.741050 4677 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.741087 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:19.741069655 +0000 UTC m=+1170.487402110 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "metrics-server-cert" not found Dec 03 14:06:11 crc kubenswrapper[4677]: E1203 14:06:11.741106 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs podName:f5d376fc-835d-44d7-954a-f838a22f7aa8 nodeName:}" failed. No retries permitted until 2025-12-03 14:06:19.741098626 +0000 UTC m=+1170.487431081 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs") pod "openstack-operator-controller-manager-775cffb9d6-lk7pd" (UID: "f5d376fc-835d-44d7-954a-f838a22f7aa8") : secret "webhook-server-cert" not found Dec 03 14:06:16 crc kubenswrapper[4677]: E1203 14:06:16.896526 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 03 14:06:16 crc kubenswrapper[4677]: E1203 14:06:16.897112 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbcfh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-8rll8_openstack-operators(226bed84-3bd3-4f3b-ad1a-ab8030a71b34): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:16 crc kubenswrapper[4677]: E1203 14:06:16.971016 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/openstack-k8s-operators/watcher-operator:d23b8876e1bcf18983498fca8ec9314bc8124a8c" Dec 03 14:06:16 crc kubenswrapper[4677]: E1203 14:06:16.971924 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/openstack-k8s-operators/watcher-operator:d23b8876e1bcf18983498fca8ec9314bc8124a8c" Dec 03 14:06:16 crc kubenswrapper[4677]: E1203 14:06:16.972194 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.129.56.48:5001/openstack-k8s-operators/watcher-operator:d23b8876e1bcf18983498fca8ec9314bc8124a8c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8t9mn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6757bc7c76-rqljj_openstack-operators(40fcffeb-3d82-431f-a208-c06bc3f30557): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:18 crc kubenswrapper[4677]: E1203 14:06:18.282540 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 14:06:18 crc kubenswrapper[4677]: E1203 14:06:18.282746 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m97j7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-vk8s9_openstack-operators(7270260d-7534-4b53-be80-69fbee93fdb7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:18 crc kubenswrapper[4677]: I1203 14:06:18.975278 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" event={"ID":"df1820a3-7030-449d-ad5d-c8bbdb75ad28","Type":"ContainerStarted","Data":"fbab0e8bef13ca695049b6841260cf6e12d72dc2d219d92987fd74355ee598ec"} Dec 03 14:06:18 crc kubenswrapper[4677]: I1203 14:06:18.979191 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" event={"ID":"01e63cef-045a-42e0-8776-2f07b1187a40","Type":"ContainerStarted","Data":"c06768c550e7877666b788cf62dca604e2645de6329bf0401c2f23dda315458c"} Dec 03 14:06:18 crc kubenswrapper[4677]: I1203 14:06:18.980657 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" event={"ID":"460e4110-77d5-476e-88d3-d9ccec539f98","Type":"ContainerStarted","Data":"dcd3afe8316ade59fac5f7187529fdf74a93c64f7a9f7da4c8783e7cbaf15982"} Dec 03 14:06:18 crc kubenswrapper[4677]: I1203 14:06:18.982111 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" event={"ID":"5236b4d8-b675-4e53-9c7c-c33606436dff","Type":"ContainerStarted","Data":"59258f891e0cac6b63d473d3c5c3c85819831bb88da501c3bd3ecd2388e32690"} Dec 03 14:06:18 crc kubenswrapper[4677]: I1203 14:06:18.982929 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" event={"ID":"8d9537f6-e3da-4acb-af76-4ad3bbc403ab","Type":"ContainerStarted","Data":"f3b1dac967c7a9f50a11ed22efa68324d3cbdb78e484d97c6b2245bdbfa4b40a"} Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.339449 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.345727 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b6f5612-8a33-46f2-8f4e-415d5bc2e807-cert\") pod \"infra-operator-controller-manager-57548d458d-qhdx7\" (UID: \"5b6f5612-8a33-46f2-8f4e-415d5bc2e807\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.346097 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.441657 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.451275 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/de88ce12-43f1-4e18-ad73-f7d2c222c4ca-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk\" (UID: \"de88ce12-43f1-4e18-ad73-f7d2c222c4ca\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.740291 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.748938 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.749064 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.758100 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-metrics-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.758495 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d376fc-835d-44d7-954a-f838a22f7aa8-webhook-certs\") pod \"openstack-operator-controller-manager-775cffb9d6-lk7pd\" (UID: \"f5d376fc-835d-44d7-954a-f838a22f7aa8\") " pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.790546 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.968425 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7"] Dec 03 14:06:19 crc kubenswrapper[4677]: I1203 14:06:19.997127 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" event={"ID":"59757e80-419c-40f9-9ab2-f6dbeff58f7c","Type":"ContainerStarted","Data":"eb83d118b0281bbb81d291ec63b5e286cc877e329e86b2d622a3155bd5794ba7"} Dec 03 14:06:20 crc kubenswrapper[4677]: I1203 14:06:20.003198 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" event={"ID":"cca481a1-d874-446a-9f63-926247653192","Type":"ContainerStarted","Data":"b0b27241762e21a5b5d7d5cb465ba27e03d9e1bf344db5f935746108044e3354"} Dec 03 14:06:20 crc kubenswrapper[4677]: I1203 14:06:20.010406 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" event={"ID":"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18","Type":"ContainerStarted","Data":"17bed488d40ef0ceeb4f3650e4bffe5bc797b3eeffb42dc55c1985a1cfe4e1dc"} Dec 03 14:06:20 crc kubenswrapper[4677]: I1203 14:06:20.014126 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" event={"ID":"829c01d5-d4d0-40a4-9513-0e60b322d4d0","Type":"ContainerStarted","Data":"591acea476ba57428d50f4c130889d853d795cc53560dc35d8862573a21baa5f"} Dec 03 14:06:20 crc kubenswrapper[4677]: I1203 14:06:20.731725 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd"] Dec 03 14:06:21 crc kubenswrapper[4677]: I1203 14:06:21.031243 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" event={"ID":"1e0b7fb1-cdb4-44a9-9508-4939f1038023","Type":"ContainerStarted","Data":"27d73d3d7f548d8436a2cb74a7b1f335d7189963952ef0f4c479a2090575368a"} Dec 03 14:06:21 crc kubenswrapper[4677]: I1203 14:06:21.033711 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" event={"ID":"5b6f5612-8a33-46f2-8f4e-415d5bc2e807","Type":"ContainerStarted","Data":"00115b82abdb95482317ee198a9b958b549bbc40a19ce6bb58791bd18f1f305e"} Dec 03 14:06:22 crc kubenswrapper[4677]: I1203 14:06:22.071903 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" event={"ID":"f5d376fc-835d-44d7-954a-f838a22f7aa8","Type":"ContainerStarted","Data":"1b5fd321cb538290a11e2961fd267282340ef8de13300c944e9c3c5403ef352b"} Dec 03 14:06:22 crc kubenswrapper[4677]: I1203 14:06:22.072728 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:22 crc kubenswrapper[4677]: I1203 14:06:22.072746 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" event={"ID":"f5d376fc-835d-44d7-954a-f838a22f7aa8","Type":"ContainerStarted","Data":"eaf85760f930eaedff45d0a1cd4427e12a00d65dade8f1d6bfc60be4aceda65d"} Dec 03 14:06:22 crc kubenswrapper[4677]: E1203 14:06:22.082687 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" podUID="226bed84-3bd3-4f3b-ad1a-ab8030a71b34" Dec 03 14:06:22 crc kubenswrapper[4677]: I1203 14:06:22.106036 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" podStartSLOduration=19.106015505 podStartE2EDuration="19.106015505s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:06:22.098071542 +0000 UTC m=+1172.844404007" watchObservedRunningTime="2025-12-03 14:06:22.106015505 +0000 UTC m=+1172.852347980" Dec 03 14:06:22 crc kubenswrapper[4677]: I1203 14:06:22.158636 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk"] Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.135705 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" event={"ID":"226bed84-3bd3-4f3b-ad1a-ab8030a71b34","Type":"ContainerStarted","Data":"ff3867969268d3c2f2bc0542b5f22040020ac693ebf4b91c69b5b093c3f019b7"} Dec 03 14:06:23 crc kubenswrapper[4677]: E1203 14:06:23.138201 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" podUID="226bed84-3bd3-4f3b-ad1a-ab8030a71b34" Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.143476 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" event={"ID":"cca481a1-d874-446a-9f63-926247653192","Type":"ContainerStarted","Data":"517e305067e11125e3548a0b8013a601ef9e7f4971000ff338fcd75e00362776"} Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.143588 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.154840 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" event={"ID":"df1820a3-7030-449d-ad5d-c8bbdb75ad28","Type":"ContainerStarted","Data":"2c7fe4928fb7f49f9c61281dddbf211ec9837f0d7f5d306b4de063ff904e1eb8"} Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.155153 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.163367 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.174165 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" podStartSLOduration=3.034107564 podStartE2EDuration="20.174151047s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.70408636 +0000 UTC m=+1155.450418815" lastFinishedPulling="2025-12-03 14:06:21.844129843 +0000 UTC m=+1172.590462298" observedRunningTime="2025-12-03 14:06:23.173638605 +0000 UTC m=+1173.919971090" watchObservedRunningTime="2025-12-03 14:06:23.174151047 +0000 UTC m=+1173.920483502" Dec 03 14:06:23 crc kubenswrapper[4677]: I1203 14:06:23.207213 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" podStartSLOduration=3.902512574 podStartE2EDuration="21.207186408s" podCreationTimestamp="2025-12-03 14:06:02 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.557788641 +0000 UTC m=+1155.304121096" lastFinishedPulling="2025-12-03 14:06:21.862462475 +0000 UTC m=+1172.608794930" observedRunningTime="2025-12-03 14:06:23.205034178 +0000 UTC m=+1173.951366643" watchObservedRunningTime="2025-12-03 14:06:23.207186408 +0000 UTC m=+1173.953518863" Dec 03 14:06:24 crc kubenswrapper[4677]: E1203 14:06:24.165780 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" podUID="226bed84-3bd3-4f3b-ad1a-ab8030a71b34" Dec 03 14:06:24 crc kubenswrapper[4677]: I1203 14:06:24.165876 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:06:28 crc kubenswrapper[4677]: I1203 14:06:28.209439 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" event={"ID":"de88ce12-43f1-4e18-ad73-f7d2c222c4ca","Type":"ContainerStarted","Data":"c4bf1bbc73b2dd9b8bb5598269d2c17f29b17c56926813dde78eabc790ad85a5"} Dec 03 14:06:29 crc kubenswrapper[4677]: I1203 14:06:29.800139 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:06:38 crc kubenswrapper[4677]: I1203 14:06:38.437001 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:06:38 crc kubenswrapper[4677]: I1203 14:06:38.437689 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:06:38 crc kubenswrapper[4677]: I1203 14:06:38.437747 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:06:38 crc kubenswrapper[4677]: I1203 14:06:38.438481 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e1aec0429580d81a4c5a21902c4598063a10937c54a8f7496939174766a6f58"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:06:38 crc kubenswrapper[4677]: I1203 14:06:38.438544 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://2e1aec0429580d81a4c5a21902c4598063a10937c54a8f7496939174766a6f58" gracePeriod=600 Dec 03 14:06:40 crc kubenswrapper[4677]: I1203 14:06:40.303873 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="2e1aec0429580d81a4c5a21902c4598063a10937c54a8f7496939174766a6f58" exitCode=0 Dec 03 14:06:40 crc kubenswrapper[4677]: I1203 14:06:40.303912 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"2e1aec0429580d81a4c5a21902c4598063a10937c54a8f7496939174766a6f58"} Dec 03 14:06:40 crc kubenswrapper[4677]: I1203 14:06:40.303939 4677 scope.go:117] "RemoveContainer" containerID="1186ac74eab5d1036a6f0e4e1fb80c6c8725f92e6867eb61894f1292d51d689c" Dec 03 14:06:41 crc kubenswrapper[4677]: E1203 14:06:41.854845 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 03 14:06:41 crc kubenswrapper[4677]: E1203 14:06:41.855401 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tdxhd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-hfwgd_openstack-operators(da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:44 crc kubenswrapper[4677]: E1203 14:06:44.044550 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:06:44 crc kubenswrapper[4677]: E1203 14:06:44.044722 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2j7kj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-45txl_openstack-operators(5236b4d8-b675-4e53-9c7c-c33606436dff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:44 crc kubenswrapper[4677]: E1203 14:06:44.046216 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" podUID="5236b4d8-b675-4e53-9c7c-c33606436dff" Dec 03 14:06:44 crc kubenswrapper[4677]: E1203 14:06:44.054400 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 14:06:44 crc kubenswrapper[4677]: E1203 14:06:44.054616 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n5clr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-5f5nf_openstack-operators(af346d56-cdf6-408d-87fc-6431b12a9cd8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:44 crc kubenswrapper[4677]: I1203 14:06:44.333762 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" event={"ID":"460e4110-77d5-476e-88d3-d9ccec539f98","Type":"ContainerStarted","Data":"657ec0752485114de8c58e27a7b166f9a62573a408a8bc35c033d03beddb84a1"} Dec 03 14:06:44 crc kubenswrapper[4677]: I1203 14:06:44.334399 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:06:44 crc kubenswrapper[4677]: I1203 14:06:44.336313 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:06:44 crc kubenswrapper[4677]: I1203 14:06:44.385269 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" podStartSLOduration=25.373593107 podStartE2EDuration="42.385231123s" podCreationTimestamp="2025-12-03 14:06:02 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.880549793 +0000 UTC m=+1155.626882248" lastFinishedPulling="2025-12-03 14:06:21.892187819 +0000 UTC m=+1172.638520264" observedRunningTime="2025-12-03 14:06:44.374375542 +0000 UTC m=+1195.120708017" watchObservedRunningTime="2025-12-03 14:06:44.385231123 +0000 UTC m=+1195.131563588" Dec 03 14:06:44 crc kubenswrapper[4677]: E1203 14:06:44.965501 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 03 14:06:44 crc kubenswrapper[4677]: E1203 14:06:44.965723 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qnc2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-8jjpl_openstack-operators(bbb9ba89-c82a-4cb6-8346-a76d2f24fd38): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:45 crc kubenswrapper[4677]: I1203 14:06:45.341016 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:06:45 crc kubenswrapper[4677]: I1203 14:06:45.343719 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:06:45 crc kubenswrapper[4677]: E1203 14:06:45.495774 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 03 14:06:45 crc kubenswrapper[4677]: E1203 14:06:45.495993 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4bjr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-jmxdr_openstack-operators(85cba130-9ba2-48b7-917f-14946bc72419): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:06:46 crc kubenswrapper[4677]: E1203 14:06:46.813191 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podUID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" Dec 03 14:06:46 crc kubenswrapper[4677]: E1203 14:06:46.896878 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" podUID="40fcffeb-3d82-431f-a208-c06bc3f30557" Dec 03 14:06:46 crc kubenswrapper[4677]: E1203 14:06:46.998931 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" Dec 03 14:06:47 crc kubenswrapper[4677]: E1203 14:06:47.117100 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" podUID="7270260d-7534-4b53-be80-69fbee93fdb7" Dec 03 14:06:47 crc kubenswrapper[4677]: E1203 14:06:47.175860 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" Dec 03 14:06:47 crc kubenswrapper[4677]: E1203 14:06:47.248673 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.403778 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" event={"ID":"40fcffeb-3d82-431f-a208-c06bc3f30557","Type":"ContainerStarted","Data":"de4ad695056e36771c8bcc09f5b10937ead9ea176b6a5cd9fc40105dc8526941"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.407638 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" event={"ID":"5b6f5612-8a33-46f2-8f4e-415d5bc2e807","Type":"ContainerStarted","Data":"41f627a7c6d49357b79610ee5311bc324802dbeb807e28d94926fd48d1a97da3"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.407670 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" event={"ID":"5b6f5612-8a33-46f2-8f4e-415d5bc2e807","Type":"ContainerStarted","Data":"dde233158412629ff036ebfb0445934e4eb70f069399d97f2472ae5a5f015886"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.408276 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.415908 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" event={"ID":"8d9537f6-e3da-4acb-af76-4ad3bbc403ab","Type":"ContainerStarted","Data":"c18f39ebc70695f3fa407779034e40e85072c5070a1d6c3d096ed9a008378e10"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.422054 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" event={"ID":"882195ac-f224-4563-b00d-0291e40f3204","Type":"ContainerStarted","Data":"956c9e64e809de334bb8047b080d2bb2bbcae4bccfdb1ee384aa8c2a4a6d44ca"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.422097 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" event={"ID":"882195ac-f224-4563-b00d-0291e40f3204","Type":"ContainerStarted","Data":"bb386a05b822f2476f32b947d5b8373fab71333b3ddc12a3ad8cdf3ce00422ac"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.422320 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.429457 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" event={"ID":"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18","Type":"ContainerStarted","Data":"c178323a2b7645517544f6301e5a1f95d46c49d44a98dcf858a8cc8c42b97f0c"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.429712 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.432175 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.435772 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" event={"ID":"01e63cef-045a-42e0-8776-2f07b1187a40","Type":"ContainerStarted","Data":"7cc094db1e036911df214e218b7813b15b54378201af4598199bb41fdc3073d2"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.435983 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.438485 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" event={"ID":"85cba130-9ba2-48b7-917f-14946bc72419","Type":"ContainerStarted","Data":"8ef02da22145ad1265f09ee9644ee47febdf1c027530d9201913f64ae9ef9eb0"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.439441 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:06:47 crc kubenswrapper[4677]: E1203 14:06:47.440074 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.444570 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" event={"ID":"47d9316a-afea-441d-8c42-45e90efe4d5a","Type":"ContainerStarted","Data":"32345eb607b0c9214179f64cb23f4281c7f37f14b7c74397e37fe4d93945d3ea"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.444613 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" event={"ID":"47d9316a-afea-441d-8c42-45e90efe4d5a","Type":"ContainerStarted","Data":"c8c0dd542f85475071f120804d614a85b0b6e757ac4b9e1dcf43d9e5df871d3f"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.445288 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.466361 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" event={"ID":"1e0b7fb1-cdb4-44a9-9508-4939f1038023","Type":"ContainerStarted","Data":"bd55d0e0dc18cf33812064fd74795d0263009a9c44a366ead1fede86dac2d998"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.467427 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.473029 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podStartSLOduration=4.620182843 podStartE2EDuration="44.473013571s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.123024707 +0000 UTC m=+1155.869357152" lastFinishedPulling="2025-12-03 14:06:44.975855385 +0000 UTC m=+1195.722187880" observedRunningTime="2025-12-03 14:06:47.471105677 +0000 UTC m=+1198.217438132" watchObservedRunningTime="2025-12-03 14:06:47.473013571 +0000 UTC m=+1198.219346026" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.473238 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.500501 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" podStartSLOduration=28.385923349 podStartE2EDuration="45.500483574s" podCreationTimestamp="2025-12-03 14:06:02 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.777649375 +0000 UTC m=+1155.523981830" lastFinishedPulling="2025-12-03 14:06:21.8922096 +0000 UTC m=+1172.638542055" observedRunningTime="2025-12-03 14:06:47.499940432 +0000 UTC m=+1198.246272897" watchObservedRunningTime="2025-12-03 14:06:47.500483574 +0000 UTC m=+1198.246816039" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.506425 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" event={"ID":"af346d56-cdf6-408d-87fc-6431b12a9cd8","Type":"ContainerStarted","Data":"32094e5fb6f9aa4ebbb933ca3709430ff879e98c2e811179557b96d4af569b0b"} Dec 03 14:06:47 crc kubenswrapper[4677]: E1203 14:06:47.509875 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.511365 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" event={"ID":"27ee4b91-853f-444b-8663-d913a1d9d1e5","Type":"ContainerStarted","Data":"e32dd67e5f14673801555948e88d4ec16dc84039ae06ae2447aec42ad600fcc4"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.535307 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" podStartSLOduration=20.634000995 podStartE2EDuration="45.535287595s" podCreationTimestamp="2025-12-03 14:06:02 +0000 UTC" firstStartedPulling="2025-12-03 14:06:20.074613377 +0000 UTC m=+1170.820945822" lastFinishedPulling="2025-12-03 14:06:44.975899967 +0000 UTC m=+1195.722232422" observedRunningTime="2025-12-03 14:06:47.518066739 +0000 UTC m=+1198.264399214" watchObservedRunningTime="2025-12-03 14:06:47.535287595 +0000 UTC m=+1198.281620060" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.552816 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" event={"ID":"59757e80-419c-40f9-9ab2-f6dbeff58f7c","Type":"ContainerStarted","Data":"9950418285f4c5fcccf8e7094a63b26c8b0f3611d0868f1cde0da1cd4eed85f4"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.554871 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.559442 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" podStartSLOduration=3.295905625 podStartE2EDuration="44.559423702s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.874080136 +0000 UTC m=+1155.620412591" lastFinishedPulling="2025-12-03 14:06:46.137598213 +0000 UTC m=+1196.883930668" observedRunningTime="2025-12-03 14:06:47.558328696 +0000 UTC m=+1198.304661161" watchObservedRunningTime="2025-12-03 14:06:47.559423702 +0000 UTC m=+1198.305756157" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.575574 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" event={"ID":"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38","Type":"ContainerStarted","Data":"d1da0b286b8982402874e2b58d996ee4674f474455b795ad8fc6be1b0d6bb5cd"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.585333 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:06:47 crc kubenswrapper[4677]: E1203 14:06:47.587215 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.589239 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"960b2bbd5800fa0b1e3c63b4bd748d2c40afb2ceea7aa3529320dbcc66c17398"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.591185 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" event={"ID":"de88ce12-43f1-4e18-ad73-f7d2c222c4ca","Type":"ContainerStarted","Data":"a146e0e11a47e91d0ccdc48f17d27d2afbd101bb43930cd8f87ad57549e8d359"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.591212 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" event={"ID":"de88ce12-43f1-4e18-ad73-f7d2c222c4ca","Type":"ContainerStarted","Data":"5bdb7e20ec10b6d3c8af7be02a54a8b50343b31898ee9fe83d955fa68ee6412c"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.591657 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.593595 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" event={"ID":"7270260d-7534-4b53-be80-69fbee93fdb7","Type":"ContainerStarted","Data":"8794da969f7235cb5ccc92070595dfcfdebc827aadb3b457054a666cb20e9a38"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.605072 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" podStartSLOduration=4.096235054 podStartE2EDuration="45.605054283s" podCreationTimestamp="2025-12-03 14:06:02 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.750291204 +0000 UTC m=+1155.496623659" lastFinishedPulling="2025-12-03 14:06:46.259110433 +0000 UTC m=+1197.005442888" observedRunningTime="2025-12-03 14:06:47.603842905 +0000 UTC m=+1198.350175370" watchObservedRunningTime="2025-12-03 14:06:47.605054283 +0000 UTC m=+1198.351386748" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.612135 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" event={"ID":"829c01d5-d4d0-40a4-9513-0e60b322d4d0","Type":"ContainerStarted","Data":"8cf874279ae9f887375f10fa09609001921d5cc32b4cd108fedb092413cc0257"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.612622 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.633195 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.636352 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" event={"ID":"226bed84-3bd3-4f3b-ad1a-ab8030a71b34","Type":"ContainerStarted","Data":"8a5c872be7aa850a9a5dbc1717b8c0b496a4d8b4785c8ab36f144e9076da670b"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.637402 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.664581 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" podStartSLOduration=4.29722672 podStartE2EDuration="44.664557993s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.11769573 +0000 UTC m=+1155.864028185" lastFinishedPulling="2025-12-03 14:06:45.485027003 +0000 UTC m=+1196.231359458" observedRunningTime="2025-12-03 14:06:47.642665448 +0000 UTC m=+1198.388997913" watchObservedRunningTime="2025-12-03 14:06:47.664557993 +0000 UTC m=+1198.410890448" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.667317 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" event={"ID":"5236b4d8-b675-4e53-9c7c-c33606436dff","Type":"ContainerStarted","Data":"b75013378bc0ce9cc22ea2220dfd51a84058f833270a5064983f8a181bfca132"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.684109 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" event={"ID":"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6","Type":"ContainerStarted","Data":"e98785a859bd2e3b348d7b3c418118bad00894c9b371efae9bf2a8521b1f743e"} Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.684704 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" podStartSLOduration=3.641002104 podStartE2EDuration="44.684693766s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.117547386 +0000 UTC m=+1155.863879841" lastFinishedPulling="2025-12-03 14:06:46.161239048 +0000 UTC m=+1196.907571503" observedRunningTime="2025-12-03 14:06:47.680550721 +0000 UTC m=+1198.426883186" watchObservedRunningTime="2025-12-03 14:06:47.684693766 +0000 UTC m=+1198.431026221" Dec 03 14:06:47 crc kubenswrapper[4677]: E1203 14:06:47.688976 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podUID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.767463 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" podStartSLOduration=3.620992577 podStartE2EDuration="44.767442063s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.060566896 +0000 UTC m=+1155.806899351" lastFinishedPulling="2025-12-03 14:06:46.207016382 +0000 UTC m=+1196.953348837" observedRunningTime="2025-12-03 14:06:47.761267381 +0000 UTC m=+1198.507599836" watchObservedRunningTime="2025-12-03 14:06:47.767442063 +0000 UTC m=+1198.513774528" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.956300 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" podStartSLOduration=26.760579796000002 podStartE2EDuration="44.956275971s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:28.068408772 +0000 UTC m=+1178.814741237" lastFinishedPulling="2025-12-03 14:06:46.264104957 +0000 UTC m=+1197.010437412" observedRunningTime="2025-12-03 14:06:47.951636445 +0000 UTC m=+1198.697968900" watchObservedRunningTime="2025-12-03 14:06:47.956275971 +0000 UTC m=+1198.702608426" Dec 03 14:06:47 crc kubenswrapper[4677]: I1203 14:06:47.981052 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" podStartSLOduration=28.142368529 podStartE2EDuration="44.981036502s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.053552017 +0000 UTC m=+1155.799884472" lastFinishedPulling="2025-12-03 14:06:21.89221999 +0000 UTC m=+1172.638552445" observedRunningTime="2025-12-03 14:06:47.97746978 +0000 UTC m=+1198.723802245" watchObservedRunningTime="2025-12-03 14:06:47.981036502 +0000 UTC m=+1198.727368967" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.095448 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" podStartSLOduration=4.45136532 podStartE2EDuration="46.095423907s" podCreationTimestamp="2025-12-03 14:06:02 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.58312103 +0000 UTC m=+1155.329453485" lastFinishedPulling="2025-12-03 14:06:46.227179617 +0000 UTC m=+1196.973512072" observedRunningTime="2025-12-03 14:06:48.041788081 +0000 UTC m=+1198.788120536" watchObservedRunningTime="2025-12-03 14:06:48.095423907 +0000 UTC m=+1198.841756362" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.105797 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" podStartSLOduration=5.787133974 podStartE2EDuration="45.105779115s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.877873914 +0000 UTC m=+1155.624206369" lastFinishedPulling="2025-12-03 14:06:44.196519055 +0000 UTC m=+1194.942851510" observedRunningTime="2025-12-03 14:06:48.102363467 +0000 UTC m=+1198.848695932" watchObservedRunningTime="2025-12-03 14:06:48.105779115 +0000 UTC m=+1198.852111570" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.693207 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" event={"ID":"7270260d-7534-4b53-be80-69fbee93fdb7","Type":"ContainerStarted","Data":"443180e9f3c2c6acf66542cfa8dea831a10053187d12b4c593c7a91a674ba7db"} Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.694467 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.696391 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" event={"ID":"40fcffeb-3d82-431f-a208-c06bc3f30557","Type":"ContainerStarted","Data":"e253c6b10b9c6a6c7b28a547eb9d90c3966ddc5e9c7e3b76e9877a8383d0b3fc"} Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.698781 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.701480 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.717924 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" podStartSLOduration=2.491916222 podStartE2EDuration="45.717903494s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.880105411 +0000 UTC m=+1155.626437866" lastFinishedPulling="2025-12-03 14:06:48.106092683 +0000 UTC m=+1198.852425138" observedRunningTime="2025-12-03 14:06:48.717220238 +0000 UTC m=+1199.463552703" watchObservedRunningTime="2025-12-03 14:06:48.717903494 +0000 UTC m=+1199.464235949" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.718099 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" podStartSLOduration=33.11436972 podStartE2EDuration="46.718094478s" podCreationTimestamp="2025-12-03 14:06:02 +0000 UTC" firstStartedPulling="2025-12-03 14:06:04.659417665 +0000 UTC m=+1155.405750120" lastFinishedPulling="2025-12-03 14:06:18.263142413 +0000 UTC m=+1169.009474878" observedRunningTime="2025-12-03 14:06:48.146186816 +0000 UTC m=+1198.892519271" watchObservedRunningTime="2025-12-03 14:06:48.718094478 +0000 UTC m=+1199.464426923" Dec 03 14:06:48 crc kubenswrapper[4677]: I1203 14:06:48.743823 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" podStartSLOduration=3.448849408 podStartE2EDuration="45.743801911s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.216805021 +0000 UTC m=+1155.963137476" lastFinishedPulling="2025-12-03 14:06:47.511757524 +0000 UTC m=+1198.258089979" observedRunningTime="2025-12-03 14:06:48.742300876 +0000 UTC m=+1199.488633341" watchObservedRunningTime="2025-12-03 14:06:48.743801911 +0000 UTC m=+1199.490134366" Dec 03 14:06:49 crc kubenswrapper[4677]: I1203 14:06:49.704876 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:06:53 crc kubenswrapper[4677]: I1203 14:06:53.288429 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:06:53 crc kubenswrapper[4677]: I1203 14:06:53.772929 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:06:53 crc kubenswrapper[4677]: I1203 14:06:53.860427 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:06:54 crc kubenswrapper[4677]: I1203 14:06:54.058666 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:06:54 crc kubenswrapper[4677]: I1203 14:06:54.089789 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:06:58 crc kubenswrapper[4677]: E1203 14:06:58.979897 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" Dec 03 14:06:59 crc kubenswrapper[4677]: I1203 14:06:59.356476 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:06:59 crc kubenswrapper[4677]: I1203 14:06:59.748757 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:06:59 crc kubenswrapper[4677]: E1203 14:06:59.981980 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" Dec 03 14:07:00 crc kubenswrapper[4677]: E1203 14:07:00.977735 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" Dec 03 14:07:02 crc kubenswrapper[4677]: I1203 14:07:02.807737 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" event={"ID":"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6","Type":"ContainerStarted","Data":"7dc7fe460118d9e5a9e8caf63764128596e0dbfd850b00313d334fd84ac09b7a"} Dec 03 14:07:02 crc kubenswrapper[4677]: I1203 14:07:02.808581 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:07:02 crc kubenswrapper[4677]: I1203 14:07:02.827921 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podStartSLOduration=2.302453457 podStartE2EDuration="59.82790227s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.068059308 +0000 UTC m=+1155.814391763" lastFinishedPulling="2025-12-03 14:07:02.593508081 +0000 UTC m=+1213.339840576" observedRunningTime="2025-12-03 14:07:02.824579344 +0000 UTC m=+1213.570911809" watchObservedRunningTime="2025-12-03 14:07:02.82790227 +0000 UTC m=+1213.574234735" Dec 03 14:07:12 crc kubenswrapper[4677]: I1203 14:07:12.886126 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" event={"ID":"85cba130-9ba2-48b7-917f-14946bc72419","Type":"ContainerStarted","Data":"cb3d9ca474f021a2d3fc8e8873de0a528ffa31dc3a7f0371b27a807a38cae050"} Dec 03 14:07:12 crc kubenswrapper[4677]: I1203 14:07:12.888085 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:07:12 crc kubenswrapper[4677]: I1203 14:07:12.906566 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podStartSLOduration=2.356386132 podStartE2EDuration="1m9.906547915s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.115454363 +0000 UTC m=+1155.861786818" lastFinishedPulling="2025-12-03 14:07:12.665616136 +0000 UTC m=+1223.411948601" observedRunningTime="2025-12-03 14:07:12.90153624 +0000 UTC m=+1223.647868715" watchObservedRunningTime="2025-12-03 14:07:12.906547915 +0000 UTC m=+1223.652880390" Dec 03 14:07:13 crc kubenswrapper[4677]: I1203 14:07:13.835370 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:07:13 crc kubenswrapper[4677]: I1203 14:07:13.898087 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" event={"ID":"af346d56-cdf6-408d-87fc-6431b12a9cd8","Type":"ContainerStarted","Data":"28df861f1deec24d0bfa111276d4f8479afacac38364db9c1ad137071747fe55"} Dec 03 14:07:13 crc kubenswrapper[4677]: I1203 14:07:13.898503 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:07:13 crc kubenswrapper[4677]: I1203 14:07:13.927929 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podStartSLOduration=3.240383158 podStartE2EDuration="1m10.927914849s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.094472195 +0000 UTC m=+1155.840804650" lastFinishedPulling="2025-12-03 14:07:12.782003886 +0000 UTC m=+1223.528336341" observedRunningTime="2025-12-03 14:07:13.926810264 +0000 UTC m=+1224.673142719" watchObservedRunningTime="2025-12-03 14:07:13.927914849 +0000 UTC m=+1224.674247304" Dec 03 14:07:17 crc kubenswrapper[4677]: I1203 14:07:17.938345 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" event={"ID":"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38","Type":"ContainerStarted","Data":"00e2823b3efa0dc8e6792e579911b5e272d67e9cd14177020b5755ab6d5645aa"} Dec 03 14:07:17 crc kubenswrapper[4677]: I1203 14:07:17.939061 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:07:17 crc kubenswrapper[4677]: I1203 14:07:17.976875 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podStartSLOduration=2.6809554220000003 podStartE2EDuration="1m14.976853887s" podCreationTimestamp="2025-12-03 14:06:03 +0000 UTC" firstStartedPulling="2025-12-03 14:06:05.117368481 +0000 UTC m=+1155.863700936" lastFinishedPulling="2025-12-03 14:07:17.413266956 +0000 UTC m=+1228.159599401" observedRunningTime="2025-12-03 14:07:17.967200194 +0000 UTC m=+1228.713532659" watchObservedRunningTime="2025-12-03 14:07:17.976853887 +0000 UTC m=+1228.723186352" Dec 03 14:07:23 crc kubenswrapper[4677]: I1203 14:07:23.714088 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:07:23 crc kubenswrapper[4677]: I1203 14:07:23.940356 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:07:24 crc kubenswrapper[4677]: I1203 14:07:24.039366 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.524013 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c448b6c-9vqqq"] Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.532731 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.537806 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-kjfpf" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.538032 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.538122 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.538286 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.555827 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c448b6c-9vqqq"] Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.587487 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c78b4f8d9-g82l7"] Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.591987 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.596596 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.600828 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c78b4f8d9-g82l7"] Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.682542 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftvrz\" (UniqueName: \"kubernetes.io/projected/e4a4a2cc-2807-4597-af11-c9de786c0886-kube-api-access-ftvrz\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.682679 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g88pr\" (UniqueName: \"kubernetes.io/projected/107b1815-1034-4a41-98f8-530bbd2127a6-kube-api-access-g88pr\") pod \"dnsmasq-dns-78c448b6c-9vqqq\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.682719 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-config\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.682758 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-dns-svc\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.682775 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107b1815-1034-4a41-98f8-530bbd2127a6-config\") pod \"dnsmasq-dns-78c448b6c-9vqqq\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.784464 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftvrz\" (UniqueName: \"kubernetes.io/projected/e4a4a2cc-2807-4597-af11-c9de786c0886-kube-api-access-ftvrz\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.784517 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g88pr\" (UniqueName: \"kubernetes.io/projected/107b1815-1034-4a41-98f8-530bbd2127a6-kube-api-access-g88pr\") pod \"dnsmasq-dns-78c448b6c-9vqqq\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.784549 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-config\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.784576 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-dns-svc\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.784594 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107b1815-1034-4a41-98f8-530bbd2127a6-config\") pod \"dnsmasq-dns-78c448b6c-9vqqq\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.785605 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107b1815-1034-4a41-98f8-530bbd2127a6-config\") pod \"dnsmasq-dns-78c448b6c-9vqqq\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.785629 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-config\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.785731 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-dns-svc\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.807483 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftvrz\" (UniqueName: \"kubernetes.io/projected/e4a4a2cc-2807-4597-af11-c9de786c0886-kube-api-access-ftvrz\") pod \"dnsmasq-dns-5c78b4f8d9-g82l7\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.813002 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g88pr\" (UniqueName: \"kubernetes.io/projected/107b1815-1034-4a41-98f8-530bbd2127a6-kube-api-access-g88pr\") pod \"dnsmasq-dns-78c448b6c-9vqqq\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.899275 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:07:41 crc kubenswrapper[4677]: I1203 14:07:41.915122 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:07:42 crc kubenswrapper[4677]: I1203 14:07:42.373750 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c78b4f8d9-g82l7"] Dec 03 14:07:42 crc kubenswrapper[4677]: I1203 14:07:42.459169 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c448b6c-9vqqq"] Dec 03 14:07:42 crc kubenswrapper[4677]: W1203 14:07:42.465382 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod107b1815_1034_4a41_98f8_530bbd2127a6.slice/crio-5be3ab57065cf00261b47555640024c107244481b8ebf8f804fa5505ae064476 WatchSource:0}: Error finding container 5be3ab57065cf00261b47555640024c107244481b8ebf8f804fa5505ae064476: Status 404 returned error can't find the container with id 5be3ab57065cf00261b47555640024c107244481b8ebf8f804fa5505ae064476 Dec 03 14:07:43 crc kubenswrapper[4677]: I1203 14:07:43.148599 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" event={"ID":"107b1815-1034-4a41-98f8-530bbd2127a6","Type":"ContainerStarted","Data":"5be3ab57065cf00261b47555640024c107244481b8ebf8f804fa5505ae064476"} Dec 03 14:07:43 crc kubenswrapper[4677]: I1203 14:07:43.151330 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" event={"ID":"e4a4a2cc-2807-4597-af11-c9de786c0886","Type":"ContainerStarted","Data":"f257222323de9cd5934d8c0f4122e391f4e3b5b1152b20585d29dd6ffc017166"} Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.521085 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c78b4f8d9-g82l7"] Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.554018 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-778bb459f7-vkrds"] Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.557714 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.563628 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778bb459f7-vkrds"] Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.647573 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz25z\" (UniqueName: \"kubernetes.io/projected/048d4741-5beb-457e-87b1-e3e215ac9012-kube-api-access-hz25z\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.647649 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-config\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.647726 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-dns-svc\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.748747 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-dns-svc\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.748810 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz25z\" (UniqueName: \"kubernetes.io/projected/048d4741-5beb-457e-87b1-e3e215ac9012-kube-api-access-hz25z\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.748874 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-config\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.749964 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-config\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.749971 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-dns-svc\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.776155 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz25z\" (UniqueName: \"kubernetes.io/projected/048d4741-5beb-457e-87b1-e3e215ac9012-kube-api-access-hz25z\") pod \"dnsmasq-dns-778bb459f7-vkrds\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.815711 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78c448b6c-9vqqq"] Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.841699 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dbc9586df-dwtkv"] Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.843364 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.854308 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dbc9586df-dwtkv"] Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.878352 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.951572 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-config\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.951699 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-dns-svc\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:45 crc kubenswrapper[4677]: I1203 14:07:45.951765 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t75z\" (UniqueName: \"kubernetes.io/projected/756512c9-e905-4c54-bbc3-73eb9e4b2e45-kube-api-access-8t75z\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.053224 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-dns-svc\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.053330 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t75z\" (UniqueName: \"kubernetes.io/projected/756512c9-e905-4c54-bbc3-73eb9e4b2e45-kube-api-access-8t75z\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.053359 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-config\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.054142 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-config\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.054785 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-dns-svc\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.080984 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t75z\" (UniqueName: \"kubernetes.io/projected/756512c9-e905-4c54-bbc3-73eb9e4b2e45-kube-api-access-8t75z\") pod \"dnsmasq-dns-5dbc9586df-dwtkv\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.172460 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.174011 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dbc9586df-dwtkv"] Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.209817 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85d946848f-qz6ws"] Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.249090 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85d946848f-qz6ws"] Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.249294 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.360389 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-dns-svc\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.360433 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf2ct\" (UniqueName: \"kubernetes.io/projected/70265ca5-cffd-42a9-b519-5c5eed5fc29a-kube-api-access-wf2ct\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.360505 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-config\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.461552 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-config\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.461644 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-dns-svc\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.461673 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf2ct\" (UniqueName: \"kubernetes.io/projected/70265ca5-cffd-42a9-b519-5c5eed5fc29a-kube-api-access-wf2ct\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.462488 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-config\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.462552 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-dns-svc\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.485439 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf2ct\" (UniqueName: \"kubernetes.io/projected/70265ca5-cffd-42a9-b519-5c5eed5fc29a-kube-api-access-wf2ct\") pod \"dnsmasq-dns-85d946848f-qz6ws\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.578018 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.684142 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.685837 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.691146 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vqzcn" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.691347 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.691363 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.691135 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.691605 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.691688 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.691881 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.692589 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775079 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775119 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d5638b0-3c75-4290-b721-d2b5733a0cd3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775148 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d5638b0-3c75-4290-b721-d2b5733a0cd3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775220 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775251 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775283 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775302 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775336 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775375 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775410 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.775430 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8jf9\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-kube-api-access-g8jf9\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.876814 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.876866 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.876892 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.876908 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.876937 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.876970 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.876999 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8jf9\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-kube-api-access-g8jf9\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.877016 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.877053 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.877070 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d5638b0-3c75-4290-b721-d2b5733a0cd3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.877088 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d5638b0-3c75-4290-b721-d2b5733a0cd3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.877851 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.877937 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.878257 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.878598 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.879095 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.879163 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.885212 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.885734 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.888598 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d5638b0-3c75-4290-b721-d2b5733a0cd3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.902535 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d5638b0-3c75-4290-b721-d2b5733a0cd3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:46 crc kubenswrapper[4677]: I1203 14:07:46.910778 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8jf9\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-kube-api-access-g8jf9\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.013812 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.015049 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.018057 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.018224 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.018361 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.019473 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-9s2bp" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.019656 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.019765 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.019912 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.024253 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.037839 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.079988 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080087 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080125 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080171 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr484\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-kube-api-access-lr484\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080356 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/315da922-99be-44c7-81d8-24f0cf55490a-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080433 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080586 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080621 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080643 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080662 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.080822 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/315da922-99be-44c7-81d8-24f0cf55490a-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182436 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182483 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182503 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr484\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-kube-api-access-lr484\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182532 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/315da922-99be-44c7-81d8-24f0cf55490a-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182560 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182606 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182625 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182641 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182656 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182687 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/315da922-99be-44c7-81d8-24f0cf55490a-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182710 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.182804 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.183304 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.183531 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.183698 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.183932 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.187846 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.187924 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/315da922-99be-44c7-81d8-24f0cf55490a-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.188627 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/315da922-99be-44c7-81d8-24f0cf55490a-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.189934 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/315da922-99be-44c7-81d8-24f0cf55490a-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.191579 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.206564 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.208772 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr484\" (UniqueName: \"kubernetes.io/projected/315da922-99be-44c7-81d8-24f0cf55490a-kube-api-access-lr484\") pod \"rabbitmq-notifications-server-0\" (UID: \"315da922-99be-44c7-81d8-24f0cf55490a\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.322544 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.344818 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.351493 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.355892 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.357357 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.357439 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.357572 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.357622 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.357699 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5dkwn" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.358887 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.360449 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.361543 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486213 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-config-data\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486278 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea48639b-8147-44bb-aa75-96c2ffc0921f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486316 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486428 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486529 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486584 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486622 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxgfq\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-kube-api-access-rxgfq\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486650 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea48639b-8147-44bb-aa75-96c2ffc0921f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486671 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486728 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.486779 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.588527 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.588570 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.588627 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-config-data\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.588649 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea48639b-8147-44bb-aa75-96c2ffc0921f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.588672 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.588705 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589273 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589309 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589329 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea48639b-8147-44bb-aa75-96c2ffc0921f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589344 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589360 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxgfq\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-kube-api-access-rxgfq\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589044 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589815 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-config-data\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.589836 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.590371 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.591539 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.591846 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.594628 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.594936 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea48639b-8147-44bb-aa75-96c2ffc0921f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.595274 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.602344 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea48639b-8147-44bb-aa75-96c2ffc0921f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.608528 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxgfq\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-kube-api-access-rxgfq\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.613444 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " pod="openstack/rabbitmq-server-0" Dec 03 14:07:47 crc kubenswrapper[4677]: I1203 14:07:47.704453 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.111165 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.113814 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.118386 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.119276 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.119942 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.120198 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lzqzt" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.130067 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.138532 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.216854 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.216961 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-config-data-default\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.216994 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a5ae3526-170f-474a-acce-41e5889470c8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.217037 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-kolla-config\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.217063 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.217079 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98qzn\" (UniqueName: \"kubernetes.io/projected/a5ae3526-170f-474a-acce-41e5889470c8-kube-api-access-98qzn\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.217094 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5ae3526-170f-474a-acce-41e5889470c8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.217110 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ae3526-170f-474a-acce-41e5889470c8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318099 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318180 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-config-data-default\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318219 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a5ae3526-170f-474a-acce-41e5889470c8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318257 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-kolla-config\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318292 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318312 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98qzn\" (UniqueName: \"kubernetes.io/projected/a5ae3526-170f-474a-acce-41e5889470c8-kube-api-access-98qzn\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318329 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5ae3526-170f-474a-acce-41e5889470c8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318346 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ae3526-170f-474a-acce-41e5889470c8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.318822 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.319096 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-config-data-default\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.320021 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.323803 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a5ae3526-170f-474a-acce-41e5889470c8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.326738 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a5ae3526-170f-474a-acce-41e5889470c8-kolla-config\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.329769 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ae3526-170f-474a-acce-41e5889470c8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.333079 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5ae3526-170f-474a-acce-41e5889470c8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.345786 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.354293 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98qzn\" (UniqueName: \"kubernetes.io/projected/a5ae3526-170f-474a-acce-41e5889470c8-kube-api-access-98qzn\") pod \"openstack-galera-0\" (UID: \"a5ae3526-170f-474a-acce-41e5889470c8\") " pod="openstack/openstack-galera-0" Dec 03 14:07:49 crc kubenswrapper[4677]: I1203 14:07:49.445977 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.457366 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.459653 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.463489 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.465203 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.465219 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.466135 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5tctm" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.481398 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.540451 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.540498 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.540750 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.540868 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/072771d7-84c7-4274-8a3d-194f18a09b24-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.540939 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nghcx\" (UniqueName: \"kubernetes.io/projected/072771d7-84c7-4274-8a3d-194f18a09b24-kube-api-access-nghcx\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.541134 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072771d7-84c7-4274-8a3d-194f18a09b24-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.541207 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/072771d7-84c7-4274-8a3d-194f18a09b24-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.541257 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.642682 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072771d7-84c7-4274-8a3d-194f18a09b24-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.642804 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/072771d7-84c7-4274-8a3d-194f18a09b24-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.642897 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.642984 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.643021 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.643112 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.643161 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/072771d7-84c7-4274-8a3d-194f18a09b24-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.643205 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nghcx\" (UniqueName: \"kubernetes.io/projected/072771d7-84c7-4274-8a3d-194f18a09b24-kube-api-access-nghcx\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.643686 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.643824 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.643871 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/072771d7-84c7-4274-8a3d-194f18a09b24-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.644181 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.645544 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/072771d7-84c7-4274-8a3d-194f18a09b24-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.647527 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072771d7-84c7-4274-8a3d-194f18a09b24-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.657484 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/072771d7-84c7-4274-8a3d-194f18a09b24-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.659823 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nghcx\" (UniqueName: \"kubernetes.io/projected/072771d7-84c7-4274-8a3d-194f18a09b24-kube-api-access-nghcx\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.662834 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"072771d7-84c7-4274-8a3d-194f18a09b24\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.779576 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.780553 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.782376 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.782568 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-s62x4" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.787277 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.788234 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.799320 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.847893 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/79a2ad48-62e8-4bc2-a569-b110da1c20ea-kolla-config\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.848005 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a2ad48-62e8-4bc2-a569-b110da1c20ea-memcached-tls-certs\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.848027 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59p6s\" (UniqueName: \"kubernetes.io/projected/79a2ad48-62e8-4bc2-a569-b110da1c20ea-kube-api-access-59p6s\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.848086 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a2ad48-62e8-4bc2-a569-b110da1c20ea-combined-ca-bundle\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.848181 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a2ad48-62e8-4bc2-a569-b110da1c20ea-config-data\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.949230 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a2ad48-62e8-4bc2-a569-b110da1c20ea-combined-ca-bundle\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.949310 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a2ad48-62e8-4bc2-a569-b110da1c20ea-config-data\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.949357 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/79a2ad48-62e8-4bc2-a569-b110da1c20ea-kolla-config\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.949393 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a2ad48-62e8-4bc2-a569-b110da1c20ea-memcached-tls-certs\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.949410 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59p6s\" (UniqueName: \"kubernetes.io/projected/79a2ad48-62e8-4bc2-a569-b110da1c20ea-kube-api-access-59p6s\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.950261 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/79a2ad48-62e8-4bc2-a569-b110da1c20ea-kolla-config\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.950795 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a2ad48-62e8-4bc2-a569-b110da1c20ea-config-data\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.952999 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a2ad48-62e8-4bc2-a569-b110da1c20ea-memcached-tls-certs\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.962092 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a2ad48-62e8-4bc2-a569-b110da1c20ea-combined-ca-bundle\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:50 crc kubenswrapper[4677]: I1203 14:07:50.968464 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59p6s\" (UniqueName: \"kubernetes.io/projected/79a2ad48-62e8-4bc2-a569-b110da1c20ea-kube-api-access-59p6s\") pod \"memcached-0\" (UID: \"79a2ad48-62e8-4bc2-a569-b110da1c20ea\") " pod="openstack/memcached-0" Dec 03 14:07:51 crc kubenswrapper[4677]: I1203 14:07:51.103555 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 14:07:52 crc kubenswrapper[4677]: I1203 14:07:52.688002 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:07:52 crc kubenswrapper[4677]: I1203 14:07:52.689275 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:07:52 crc kubenswrapper[4677]: I1203 14:07:52.691487 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-n56gt" Dec 03 14:07:52 crc kubenswrapper[4677]: I1203 14:07:52.701238 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:07:52 crc kubenswrapper[4677]: I1203 14:07:52.778368 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2pj9\" (UniqueName: \"kubernetes.io/projected/8afc071f-9f1a-4830-bd22-a1c291e5ae53-kube-api-access-j2pj9\") pod \"kube-state-metrics-0\" (UID: \"8afc071f-9f1a-4830-bd22-a1c291e5ae53\") " pod="openstack/kube-state-metrics-0" Dec 03 14:07:52 crc kubenswrapper[4677]: I1203 14:07:52.880021 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2pj9\" (UniqueName: \"kubernetes.io/projected/8afc071f-9f1a-4830-bd22-a1c291e5ae53-kube-api-access-j2pj9\") pod \"kube-state-metrics-0\" (UID: \"8afc071f-9f1a-4830-bd22-a1c291e5ae53\") " pod="openstack/kube-state-metrics-0" Dec 03 14:07:52 crc kubenswrapper[4677]: I1203 14:07:52.905087 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2pj9\" (UniqueName: \"kubernetes.io/projected/8afc071f-9f1a-4830-bd22-a1c291e5ae53-kube-api-access-j2pj9\") pod \"kube-state-metrics-0\" (UID: \"8afc071f-9f1a-4830-bd22-a1c291e5ae53\") " pod="openstack/kube-state-metrics-0" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.006486 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.966848 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.977537 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.980814 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mq9g8" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.981047 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.981367 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.982763 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.988153 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.988476 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 14:07:53 crc kubenswrapper[4677]: I1203 14:07:53.991985 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098114 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098184 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098222 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/db9e192a-61af-41fa-882c-81879f2a556a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098261 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098284 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-config\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098358 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w222\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-kube-api-access-6w222\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098672 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.098800 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/db9e192a-61af-41fa-882c-81879f2a556a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200067 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w222\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-kube-api-access-6w222\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200123 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200162 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/db9e192a-61af-41fa-882c-81879f2a556a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200199 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200227 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200250 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/db9e192a-61af-41fa-882c-81879f2a556a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200299 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.200314 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-config\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.204021 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-config\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.206873 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.219187 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/db9e192a-61af-41fa-882c-81879f2a556a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.220426 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/db9e192a-61af-41fa-882c-81879f2a556a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.223744 4677 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.223800 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b3062aca7e122ea7b9aa119622cdce8d97263fccd877be8c6301bd417d0b47b0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.224739 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w222\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-kube-api-access-6w222\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.225758 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.226770 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.266398 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:54 crc kubenswrapper[4677]: I1203 14:07:54.330040 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.898555 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jpkmm"] Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.899887 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.901250 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-qr8sk" Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.902814 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.904816 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-msfg2"] Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.906578 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.909027 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.921376 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jpkmm"] Dec 03 14:07:55 crc kubenswrapper[4677]: I1203 14:07:55.928920 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-msfg2"] Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029242 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-etc-ovs\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029514 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-run\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029556 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jvjg\" (UniqueName: \"kubernetes.io/projected/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-kube-api-access-4jvjg\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029583 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-scripts\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029602 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-lib\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029621 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-combined-ca-bundle\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029664 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-log\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029685 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-log-ovn\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029703 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-run-ovn\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029727 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-ovn-controller-tls-certs\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029781 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-run\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029797 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d11b3a89-22f0-4768-bc37-701a49737694-scripts\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.029816 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfc8k\" (UniqueName: \"kubernetes.io/projected/d11b3a89-22f0-4768-bc37-701a49737694-kube-api-access-sfc8k\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130715 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-run\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130777 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jvjg\" (UniqueName: \"kubernetes.io/projected/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-kube-api-access-4jvjg\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130802 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-scripts\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130822 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-lib\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130843 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-combined-ca-bundle\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130877 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-log\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130893 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-log-ovn\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130918 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-run-ovn\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.130966 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-ovn-controller-tls-certs\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131042 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-run\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131063 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d11b3a89-22f0-4768-bc37-701a49737694-scripts\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131080 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfc8k\" (UniqueName: \"kubernetes.io/projected/d11b3a89-22f0-4768-bc37-701a49737694-kube-api-access-sfc8k\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131101 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-etc-ovs\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131511 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-etc-ovs\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131688 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-log-ovn\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131734 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-lib\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.131734 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-log\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.132716 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d11b3a89-22f0-4768-bc37-701a49737694-var-run\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.132774 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-run\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.132864 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-var-run-ovn\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.133643 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d11b3a89-22f0-4768-bc37-701a49737694-scripts\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.133920 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-scripts\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.137878 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-combined-ca-bundle\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.138804 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-ovn-controller-tls-certs\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.149581 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jvjg\" (UniqueName: \"kubernetes.io/projected/f14465fa-42ae-4e85-bb6f-f2ef1bc5b681-kube-api-access-4jvjg\") pod \"ovn-controller-jpkmm\" (UID: \"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681\") " pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.151686 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfc8k\" (UniqueName: \"kubernetes.io/projected/d11b3a89-22f0-4768-bc37-701a49737694-kube-api-access-sfc8k\") pod \"ovn-controller-ovs-msfg2\" (UID: \"d11b3a89-22f0-4768-bc37-701a49737694\") " pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.218408 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jpkmm" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.227351 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.812496 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.815370 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.817815 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.817832 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rknbz" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.817876 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.818728 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.819125 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.823355 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.943016 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62024578-ee41-4839-9846-9a76d8785d6e-config\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.943148 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.943343 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62024578-ee41-4839-9846-9a76d8785d6e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.943423 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.943539 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.943613 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62024578-ee41-4839-9846-9a76d8785d6e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.944034 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:56 crc kubenswrapper[4677]: I1203 14:07:56.944075 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqr87\" (UniqueName: \"kubernetes.io/projected/62024578-ee41-4839-9846-9a76d8785d6e-kube-api-access-bqr87\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.045778 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62024578-ee41-4839-9846-9a76d8785d6e-config\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.045835 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.045872 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62024578-ee41-4839-9846-9a76d8785d6e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.045897 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.045930 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.046701 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62024578-ee41-4839-9846-9a76d8785d6e-config\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.046712 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62024578-ee41-4839-9846-9a76d8785d6e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.046847 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqr87\" (UniqueName: \"kubernetes.io/projected/62024578-ee41-4839-9846-9a76d8785d6e-kube-api-access-bqr87\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.046866 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.047530 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62024578-ee41-4839-9846-9a76d8785d6e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.047530 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.047997 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62024578-ee41-4839-9846-9a76d8785d6e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.051671 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.061247 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.063225 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/62024578-ee41-4839-9846-9a76d8785d6e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.100624 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqr87\" (UniqueName: \"kubernetes.io/projected/62024578-ee41-4839-9846-9a76d8785d6e-kube-api-access-bqr87\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.106062 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"62024578-ee41-4839-9846-9a76d8785d6e\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:07:57 crc kubenswrapper[4677]: I1203 14:07:57.149156 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.259062 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.260852 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.262824 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rwq4m" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.263158 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.263485 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.263713 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.273009 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.303184 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.303562 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.303715 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h422b\" (UniqueName: \"kubernetes.io/projected/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-kube-api-access-h422b\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.303751 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.308010 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.308292 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.308485 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.308565 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-config\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411640 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411697 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-config\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411739 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411809 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411837 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h422b\" (UniqueName: \"kubernetes.io/projected/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-kube-api-access-h422b\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411862 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411920 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.411995 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.412474 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.412640 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.413577 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-config\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.414032 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.418090 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.418543 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.427736 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.436587 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h422b\" (UniqueName: \"kubernetes.io/projected/9b7e7e94-5f47-4963-acb0-f40d4ae3ca16-kube-api-access-h422b\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.443452 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:00 crc kubenswrapper[4677]: I1203 14:08:00.589754 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.381847 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.382205 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.382307 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.129.56.48:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ftvrz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5c78b4f8d9-g82l7_openstack(e4a4a2cc-2807-4597-af11-c9de786c0886): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.382359 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.382369 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.382419 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.129.56.48:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g88pr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78c448b6c-9vqqq_openstack(107b1815-1034-4a41-98f8-530bbd2127a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.383402 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" podUID="e4a4a2cc-2807-4597-af11-c9de786c0886" Dec 03 14:08:04 crc kubenswrapper[4677]: E1203 14:08:04.383448 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" podUID="107b1815-1034-4a41-98f8-530bbd2127a6" Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.069991 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: W1203 14:08:05.071802 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod315da922_99be_44c7_81d8_24f0cf55490a.slice/crio-6bde3fe915450fdb427d345734fdd05a29c4d2af85f347a1a6168bff7263e9b5 WatchSource:0}: Error finding container 6bde3fe915450fdb427d345734fdd05a29c4d2af85f347a1a6168bff7263e9b5: Status 404 returned error can't find the container with id 6bde3fe915450fdb427d345734fdd05a29c4d2af85f347a1a6168bff7263e9b5 Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.099626 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.116267 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dbc9586df-dwtkv"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.128247 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.157203 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.186776 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85d946848f-qz6ws"] Dec 03 14:08:05 crc kubenswrapper[4677]: W1203 14:08:05.189126 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5ae3526_170f_474a_acce_41e5889470c8.slice/crio-1cf8ea870dc2e64d53930958e810525b3f9abef13a5b3c07edfd941ab7f4487a WatchSource:0}: Error finding container 1cf8ea870dc2e64d53930958e810525b3f9abef13a5b3c07edfd941ab7f4487a: Status 404 returned error can't find the container with id 1cf8ea870dc2e64d53930958e810525b3f9abef13a5b3c07edfd941ab7f4487a Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.338254 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: W1203 14:08:05.352248 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod072771d7_84c7_4274_8a3d_194f18a09b24.slice/crio-52063c2524fd4655ecb2b4abac8993a0f53db77b14ce91a71bf248b96c580847 WatchSource:0}: Error finding container 52063c2524fd4655ecb2b4abac8993a0f53db77b14ce91a71bf248b96c580847: Status 404 returned error can't find the container with id 52063c2524fd4655ecb2b4abac8993a0f53db77b14ce91a71bf248b96c580847 Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.363395 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778bb459f7-vkrds"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.378743 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.451008 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jpkmm"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.454759 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"315da922-99be-44c7-81d8-24f0cf55490a","Type":"ContainerStarted","Data":"6bde3fe915450fdb427d345734fdd05a29c4d2af85f347a1a6168bff7263e9b5"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.455959 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"072771d7-84c7-4274-8a3d-194f18a09b24","Type":"ContainerStarted","Data":"52063c2524fd4655ecb2b4abac8993a0f53db77b14ce91a71bf248b96c580847"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.458259 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" event={"ID":"048d4741-5beb-457e-87b1-e3e215ac9012","Type":"ContainerStarted","Data":"f24a546a078832ce93898e1a5ba55109e6405495d6d845e1cc8ccb41a43feb84"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.459713 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a5ae3526-170f-474a-acce-41e5889470c8","Type":"ContainerStarted","Data":"1cf8ea870dc2e64d53930958e810525b3f9abef13a5b3c07edfd941ab7f4487a"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.460176 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.466748 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" event={"ID":"756512c9-e905-4c54-bbc3-73eb9e4b2e45","Type":"ContainerStarted","Data":"91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.466791 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" event={"ID":"756512c9-e905-4c54-bbc3-73eb9e4b2e45","Type":"ContainerStarted","Data":"ece826ffcd700d73739ea8402cd636b083f89c1b56bf04dd62a644c2d0ce4543"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.468805 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.475418 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"79a2ad48-62e8-4bc2-a569-b110da1c20ea","Type":"ContainerStarted","Data":"8466fcf02baa867fada0c9357f71f35fe27eef4169d43cb2e8eaeb0f111a7a21"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.477743 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7d5638b0-3c75-4290-b721-d2b5733a0cd3","Type":"ContainerStarted","Data":"1eaca61478414b41173979934f53b8d704256a03bee66f24335b8368519ea10c"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.482105 4677 generic.go:334] "Generic (PLEG): container finished" podID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerID="b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b" exitCode=0 Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.483393 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" event={"ID":"70265ca5-cffd-42a9-b519-5c5eed5fc29a","Type":"ContainerDied","Data":"b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b"} Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.484398 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" event={"ID":"70265ca5-cffd-42a9-b519-5c5eed5fc29a","Type":"ContainerStarted","Data":"83891f189a8a979639ddc22605e74721ce2c8bad511362ef7a8630adc1b37cce"} Dec 03 14:08:05 crc kubenswrapper[4677]: W1203 14:08:05.513106 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8afc071f_9f1a_4830_bd22_a1c291e5ae53.slice/crio-d3a01e25e672dea07b5760eae472232cf2e416826d46570dcb7a84499635e560 WatchSource:0}: Error finding container d3a01e25e672dea07b5760eae472232cf2e416826d46570dcb7a84499635e560: Status 404 returned error can't find the container with id d3a01e25e672dea07b5760eae472232cf2e416826d46570dcb7a84499635e560 Dec 03 14:08:05 crc kubenswrapper[4677]: W1203 14:08:05.515450 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb9e192a_61af_41fa_882c_81879f2a556a.slice/crio-2ac15fca35376f76ef2d15786dfa90bf4a4904216e15d5ecf5f79fce40752e7d WatchSource:0}: Error finding container 2ac15fca35376f76ef2d15786dfa90bf4a4904216e15d5ecf5f79fce40752e7d: Status 404 returned error can't find the container with id 2ac15fca35376f76ef2d15786dfa90bf4a4904216e15d5ecf5f79fce40752e7d Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.550609 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-msfg2"] Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.935349 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.979932 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:08:05 crc kubenswrapper[4677]: I1203 14:08:05.986311 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.028847 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t75z\" (UniqueName: \"kubernetes.io/projected/756512c9-e905-4c54-bbc3-73eb9e4b2e45-kube-api-access-8t75z\") pod \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.028910 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g88pr\" (UniqueName: \"kubernetes.io/projected/107b1815-1034-4a41-98f8-530bbd2127a6-kube-api-access-g88pr\") pod \"107b1815-1034-4a41-98f8-530bbd2127a6\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.029165 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftvrz\" (UniqueName: \"kubernetes.io/projected/e4a4a2cc-2807-4597-af11-c9de786c0886-kube-api-access-ftvrz\") pod \"e4a4a2cc-2807-4597-af11-c9de786c0886\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.029208 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107b1815-1034-4a41-98f8-530bbd2127a6-config\") pod \"107b1815-1034-4a41-98f8-530bbd2127a6\" (UID: \"107b1815-1034-4a41-98f8-530bbd2127a6\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.029336 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-config\") pod \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.029369 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-dns-svc\") pod \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\" (UID: \"756512c9-e905-4c54-bbc3-73eb9e4b2e45\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.029401 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-config\") pod \"e4a4a2cc-2807-4597-af11-c9de786c0886\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.029417 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-dns-svc\") pod \"e4a4a2cc-2807-4597-af11-c9de786c0886\" (UID: \"e4a4a2cc-2807-4597-af11-c9de786c0886\") " Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.030487 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/107b1815-1034-4a41-98f8-530bbd2127a6-config" (OuterVolumeSpecName: "config") pod "107b1815-1034-4a41-98f8-530bbd2127a6" (UID: "107b1815-1034-4a41-98f8-530bbd2127a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.030543 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-config" (OuterVolumeSpecName: "config") pod "e4a4a2cc-2807-4597-af11-c9de786c0886" (UID: "e4a4a2cc-2807-4597-af11-c9de786c0886"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.030634 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4a4a2cc-2807-4597-af11-c9de786c0886" (UID: "e4a4a2cc-2807-4597-af11-c9de786c0886"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.038437 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756512c9-e905-4c54-bbc3-73eb9e4b2e45-kube-api-access-8t75z" (OuterVolumeSpecName: "kube-api-access-8t75z") pod "756512c9-e905-4c54-bbc3-73eb9e4b2e45" (UID: "756512c9-e905-4c54-bbc3-73eb9e4b2e45"). InnerVolumeSpecName "kube-api-access-8t75z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.038470 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/107b1815-1034-4a41-98f8-530bbd2127a6-kube-api-access-g88pr" (OuterVolumeSpecName: "kube-api-access-g88pr") pod "107b1815-1034-4a41-98f8-530bbd2127a6" (UID: "107b1815-1034-4a41-98f8-530bbd2127a6"). InnerVolumeSpecName "kube-api-access-g88pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.039145 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4a4a2cc-2807-4597-af11-c9de786c0886-kube-api-access-ftvrz" (OuterVolumeSpecName: "kube-api-access-ftvrz") pod "e4a4a2cc-2807-4597-af11-c9de786c0886" (UID: "e4a4a2cc-2807-4597-af11-c9de786c0886"). InnerVolumeSpecName "kube-api-access-ftvrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.051082 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "756512c9-e905-4c54-bbc3-73eb9e4b2e45" (UID: "756512c9-e905-4c54-bbc3-73eb9e4b2e45"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.052861 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-config" (OuterVolumeSpecName: "config") pod "756512c9-e905-4c54-bbc3-73eb9e4b2e45" (UID: "756512c9-e905-4c54-bbc3-73eb9e4b2e45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131836 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131879 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/756512c9-e905-4c54-bbc3-73eb9e4b2e45-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131888 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131897 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4a4a2cc-2807-4597-af11-c9de786c0886-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131907 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t75z\" (UniqueName: \"kubernetes.io/projected/756512c9-e905-4c54-bbc3-73eb9e4b2e45-kube-api-access-8t75z\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131921 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g88pr\" (UniqueName: \"kubernetes.io/projected/107b1815-1034-4a41-98f8-530bbd2127a6-kube-api-access-g88pr\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131930 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftvrz\" (UniqueName: \"kubernetes.io/projected/e4a4a2cc-2807-4597-af11-c9de786c0886-kube-api-access-ftvrz\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.131939 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107b1815-1034-4a41-98f8-530bbd2127a6-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.442293 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.503278 4677 generic.go:334] "Generic (PLEG): container finished" podID="048d4741-5beb-457e-87b1-e3e215ac9012" containerID="74e62feded806add1645de56c1938a882edd233eb024bb5e340a73d56680437f" exitCode=0 Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.503368 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" event={"ID":"048d4741-5beb-457e-87b1-e3e215ac9012","Type":"ContainerDied","Data":"74e62feded806add1645de56c1938a882edd233eb024bb5e340a73d56680437f"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.507498 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-msfg2" event={"ID":"d11b3a89-22f0-4768-bc37-701a49737694","Type":"ContainerStarted","Data":"1eab68a9b92f09d50bec9b348283ac2f57ea09b7e92d69cb9e48b0ed842f53f3"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.510205 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerStarted","Data":"2ac15fca35376f76ef2d15786dfa90bf4a4904216e15d5ecf5f79fce40752e7d"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.511758 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" event={"ID":"e4a4a2cc-2807-4597-af11-c9de786c0886","Type":"ContainerDied","Data":"f257222323de9cd5934d8c0f4122e391f4e3b5b1152b20585d29dd6ffc017166"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.511776 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78b4f8d9-g82l7" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.515026 4677 generic.go:334] "Generic (PLEG): container finished" podID="756512c9-e905-4c54-bbc3-73eb9e4b2e45" containerID="91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619" exitCode=0 Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.515076 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" event={"ID":"756512c9-e905-4c54-bbc3-73eb9e4b2e45","Type":"ContainerDied","Data":"91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.515092 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" event={"ID":"756512c9-e905-4c54-bbc3-73eb9e4b2e45","Type":"ContainerDied","Data":"ece826ffcd700d73739ea8402cd636b083f89c1b56bf04dd62a644c2d0ce4543"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.515107 4677 scope.go:117] "RemoveContainer" containerID="91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.515214 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbc9586df-dwtkv" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.522650 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ea48639b-8147-44bb-aa75-96c2ffc0921f","Type":"ContainerStarted","Data":"64dd34abfa008f590cdb828d5463a92c8d0786d45d11451c549d0878ae5339ee"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.526088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8afc071f-9f1a-4830-bd22-a1c291e5ae53","Type":"ContainerStarted","Data":"d3a01e25e672dea07b5760eae472232cf2e416826d46570dcb7a84499635e560"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.527047 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jpkmm" event={"ID":"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681","Type":"ContainerStarted","Data":"a4427ca576f5b9e638254d058493f9d2f6dca5a3a1a811a3da5f37d3261ffce0"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.536603 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" event={"ID":"70265ca5-cffd-42a9-b519-5c5eed5fc29a","Type":"ContainerStarted","Data":"0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.536883 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.539621 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" event={"ID":"107b1815-1034-4a41-98f8-530bbd2127a6","Type":"ContainerDied","Data":"5be3ab57065cf00261b47555640024c107244481b8ebf8f804fa5505ae064476"} Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.539741 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c448b6c-9vqqq" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.541410 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.573002 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c78b4f8d9-g82l7"] Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.578628 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c78b4f8d9-g82l7"] Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.583917 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" podStartSLOduration=20.583892923 podStartE2EDuration="20.583892923s" podCreationTimestamp="2025-12-03 14:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:06.577347693 +0000 UTC m=+1277.323680148" watchObservedRunningTime="2025-12-03 14:08:06.583892923 +0000 UTC m=+1277.330225388" Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.651672 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dbc9586df-dwtkv"] Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.655964 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dbc9586df-dwtkv"] Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.671183 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78c448b6c-9vqqq"] Dec 03 14:08:06 crc kubenswrapper[4677]: I1203 14:08:06.675879 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78c448b6c-9vqqq"] Dec 03 14:08:07 crc kubenswrapper[4677]: I1203 14:08:07.547709 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16","Type":"ContainerStarted","Data":"1109619bb636c371b109117428e234bef3393e9e46795ce8955029964b8a5d26"} Dec 03 14:08:07 crc kubenswrapper[4677]: I1203 14:08:07.550572 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"62024578-ee41-4839-9846-9a76d8785d6e","Type":"ContainerStarted","Data":"e94ad9d13427313cf90be2e66d4d992fdddecb99ad9a7bbc0f9014b0d0b9ffe0"} Dec 03 14:08:07 crc kubenswrapper[4677]: I1203 14:08:07.987119 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="107b1815-1034-4a41-98f8-530bbd2127a6" path="/var/lib/kubelet/pods/107b1815-1034-4a41-98f8-530bbd2127a6/volumes" Dec 03 14:08:07 crc kubenswrapper[4677]: I1203 14:08:07.987474 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="756512c9-e905-4c54-bbc3-73eb9e4b2e45" path="/var/lib/kubelet/pods/756512c9-e905-4c54-bbc3-73eb9e4b2e45/volumes" Dec 03 14:08:07 crc kubenswrapper[4677]: I1203 14:08:07.987926 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4a4a2cc-2807-4597-af11-c9de786c0886" path="/var/lib/kubelet/pods/e4a4a2cc-2807-4597-af11-c9de786c0886/volumes" Dec 03 14:08:11 crc kubenswrapper[4677]: I1203 14:08:11.580123 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:08:11 crc kubenswrapper[4677]: I1203 14:08:11.635196 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778bb459f7-vkrds"] Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.633012 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.633456 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.633592 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nghcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(072771d7-84c7-4274-8a3d-194f18a09b24): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.635658 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="072771d7-84c7-4274-8a3d-194f18a09b24" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.673782 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="072771d7-84c7-4274-8a3d-194f18a09b24" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.692092 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.692162 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.692311 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98qzn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(a5ae3526-170f-474a-acce-41e5889470c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.693657 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="a5ae3526-170f-474a-acce-41e5889470c8" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.965090 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.965306 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init-config-reloader,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62,Command:[/bin/prometheus-config-reloader],Args:[--watch-interval=0 --listen-address=:8081 --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:reloader-init,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:SHARD,Value:0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/prometheus/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-out,ReadOnly:false,MountPath:/etc/prometheus/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-0,ReadOnly:false,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6w222,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod prometheus-metric-storage-0_openstack(db9e192a-61af-41fa-882c-81879f2a556a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.967216 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/prometheus-metric-storage-0" podUID="db9e192a-61af-41fa-882c-81879f2a556a" Dec 03 14:08:20 crc kubenswrapper[4677]: I1203 14:08:20.967277 4677 scope.go:117] "RemoveContainer" containerID="91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.967842 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619\": container with ID starting with 91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619 not found: ID does not exist" containerID="91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619" Dec 03 14:08:20 crc kubenswrapper[4677]: I1203 14:08:20.967874 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619"} err="failed to get container status \"91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619\": rpc error: code = NotFound desc = could not find container \"91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619\": container with ID starting with 91c0994c54111632dcc47d107311f7e8f76c4228fd8b97841ac1b85b14a68619 not found: ID does not exist" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.985029 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.985103 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.985278 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:38.129.56.48:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8jf9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(7d5638b0-3c75-4290-b721-d2b5733a0cd3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:08:20 crc kubenswrapper[4677]: E1203 14:08:20.986561 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.248025 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-ovn-sb-db-server:watcher_latest" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.248071 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-ovn-sb-db-server:watcher_latest" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.248200 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:38.129.56.48:5001/podified-master-centos10/openstack-ovn-sb-db-server:watcher_latest,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n669h5f9h5ddh5dfh695h549h67bhbh6hb7h589h555hc7h64fh66fh598h5d4h56ch5f8h5dbhc7h667hc6h5ddh59bh596h66bh98h5d9h685hf7h5f6q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h422b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(9b7e7e94-5f47-4963-acb0-f40d4ae3ca16): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.440148 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.440232 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.440367 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:38.129.56.48:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncbh5f9h6dh5cfh689h5f7h5bfh5bbh696h697h6ch5d9h87h547hf4h596h644hcdh6ch89hb4h545h5d7h96h5c6h557h678h6h57h57bh5dch659q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqr87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(62024578-ee41-4839-9846-9a76d8785d6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.684633 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-mariadb:watcher_latest\\\"\"" pod="openstack/openstack-galera-0" podUID="a5ae3526-170f-474a-acce-41e5889470c8" Dec 03 14:08:21 crc kubenswrapper[4677]: E1203 14:08:21.684788 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62\\\"\"" pod="openstack/prometheus-metric-storage-0" podUID="db9e192a-61af-41fa-882c-81879f2a556a" Dec 03 14:08:22 crc kubenswrapper[4677]: E1203 14:08:22.308849 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 03 14:08:22 crc kubenswrapper[4677]: E1203 14:08:22.308899 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 03 14:08:22 crc kubenswrapper[4677]: E1203 14:08:22.309071 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j2pj9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(8afc071f-9f1a-4830-bd22-a1c291e5ae53): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:08:22 crc kubenswrapper[4677]: E1203 14:08:22.311210 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" Dec 03 14:08:22 crc kubenswrapper[4677]: I1203 14:08:22.693422 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"79a2ad48-62e8-4bc2-a569-b110da1c20ea","Type":"ContainerStarted","Data":"fb50cbee7bc6df725c234942d29c117423609bcb3f1c2611a5119e4f622a625b"} Dec 03 14:08:22 crc kubenswrapper[4677]: I1203 14:08:22.694054 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 14:08:22 crc kubenswrapper[4677]: I1203 14:08:22.700321 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" event={"ID":"048d4741-5beb-457e-87b1-e3e215ac9012","Type":"ContainerStarted","Data":"97bc3554534b6a204b42c10ef282893412d0228537290d83522bf1bfa1005568"} Dec 03 14:08:22 crc kubenswrapper[4677]: I1203 14:08:22.700427 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" podUID="048d4741-5beb-457e-87b1-e3e215ac9012" containerName="dnsmasq-dns" containerID="cri-o://97bc3554534b6a204b42c10ef282893412d0228537290d83522bf1bfa1005568" gracePeriod=10 Dec 03 14:08:22 crc kubenswrapper[4677]: I1203 14:08:22.700450 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:08:22 crc kubenswrapper[4677]: E1203 14:08:22.705212 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" Dec 03 14:08:22 crc kubenswrapper[4677]: I1203 14:08:22.716538 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=16.577279171 podStartE2EDuration="32.716517566s" podCreationTimestamp="2025-12-03 14:07:50 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.108087072 +0000 UTC m=+1275.854419527" lastFinishedPulling="2025-12-03 14:08:21.247325467 +0000 UTC m=+1291.993657922" observedRunningTime="2025-12-03 14:08:22.709620267 +0000 UTC m=+1293.455952742" watchObservedRunningTime="2025-12-03 14:08:22.716517566 +0000 UTC m=+1293.462850021" Dec 03 14:08:22 crc kubenswrapper[4677]: I1203 14:08:22.765140 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" podStartSLOduration=37.765120835 podStartE2EDuration="37.765120835s" podCreationTimestamp="2025-12-03 14:07:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:22.763079998 +0000 UTC m=+1293.509412453" watchObservedRunningTime="2025-12-03 14:08:22.765120835 +0000 UTC m=+1293.511453290" Dec 03 14:08:23 crc kubenswrapper[4677]: I1203 14:08:23.716151 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jpkmm" event={"ID":"f14465fa-42ae-4e85-bb6f-f2ef1bc5b681","Type":"ContainerStarted","Data":"26429cf38bf6bae05f0202f49871aa808669ba6cd71179cb4a116a181d8d2b8e"} Dec 03 14:08:23 crc kubenswrapper[4677]: I1203 14:08:23.716527 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jpkmm" Dec 03 14:08:23 crc kubenswrapper[4677]: I1203 14:08:23.719424 4677 generic.go:334] "Generic (PLEG): container finished" podID="048d4741-5beb-457e-87b1-e3e215ac9012" containerID="97bc3554534b6a204b42c10ef282893412d0228537290d83522bf1bfa1005568" exitCode=0 Dec 03 14:08:23 crc kubenswrapper[4677]: I1203 14:08:23.719515 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" event={"ID":"048d4741-5beb-457e-87b1-e3e215ac9012","Type":"ContainerDied","Data":"97bc3554534b6a204b42c10ef282893412d0228537290d83522bf1bfa1005568"} Dec 03 14:08:23 crc kubenswrapper[4677]: I1203 14:08:23.722297 4677 generic.go:334] "Generic (PLEG): container finished" podID="d11b3a89-22f0-4768-bc37-701a49737694" containerID="ac6f7592d1d9cb8bef5eb90ea826e08f0a28c8a90753979be4c7195096369d77" exitCode=0 Dec 03 14:08:23 crc kubenswrapper[4677]: I1203 14:08:23.722337 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-msfg2" event={"ID":"d11b3a89-22f0-4768-bc37-701a49737694","Type":"ContainerDied","Data":"ac6f7592d1d9cb8bef5eb90ea826e08f0a28c8a90753979be4c7195096369d77"} Dec 03 14:08:23 crc kubenswrapper[4677]: I1203 14:08:23.738881 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jpkmm" podStartSLOduration=12.639930106 podStartE2EDuration="28.738860883s" podCreationTimestamp="2025-12-03 14:07:55 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.480657113 +0000 UTC m=+1276.226989558" lastFinishedPulling="2025-12-03 14:08:21.57958788 +0000 UTC m=+1292.325920335" observedRunningTime="2025-12-03 14:08:23.73569438 +0000 UTC m=+1294.482026835" watchObservedRunningTime="2025-12-03 14:08:23.738860883 +0000 UTC m=+1294.485193328" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.533503 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:08:24 crc kubenswrapper[4677]: E1203 14:08:24.649200 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="62024578-ee41-4839-9846-9a76d8785d6e" Dec 03 14:08:24 crc kubenswrapper[4677]: E1203 14:08:24.649526 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="9b7e7e94-5f47-4963-acb0-f40d4ae3ca16" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.688925 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-config\") pod \"048d4741-5beb-457e-87b1-e3e215ac9012\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.689042 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz25z\" (UniqueName: \"kubernetes.io/projected/048d4741-5beb-457e-87b1-e3e215ac9012-kube-api-access-hz25z\") pod \"048d4741-5beb-457e-87b1-e3e215ac9012\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.689793 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-dns-svc\") pod \"048d4741-5beb-457e-87b1-e3e215ac9012\" (UID: \"048d4741-5beb-457e-87b1-e3e215ac9012\") " Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.692906 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048d4741-5beb-457e-87b1-e3e215ac9012-kube-api-access-hz25z" (OuterVolumeSpecName: "kube-api-access-hz25z") pod "048d4741-5beb-457e-87b1-e3e215ac9012" (UID: "048d4741-5beb-457e-87b1-e3e215ac9012"). InnerVolumeSpecName "kube-api-access-hz25z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.726075 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-config" (OuterVolumeSpecName: "config") pod "048d4741-5beb-457e-87b1-e3e215ac9012" (UID: "048d4741-5beb-457e-87b1-e3e215ac9012"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.727125 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "048d4741-5beb-457e-87b1-e3e215ac9012" (UID: "048d4741-5beb-457e-87b1-e3e215ac9012"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.735115 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-msfg2" event={"ID":"d11b3a89-22f0-4768-bc37-701a49737694","Type":"ContainerStarted","Data":"37366db58f72ad6226e42a6d89ec7f5f6cfa6af844bbf8ade44fc9f06b344ce7"} Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.736516 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7d5638b0-3c75-4290-b721-d2b5733a0cd3","Type":"ContainerStarted","Data":"c6d03f88aa96a8f0d3893995f7b8086caba5f3f4e8991faae35b447b28f9527a"} Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.738303 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16","Type":"ContainerStarted","Data":"7bd442cbefe908e6fa4a457290eb33cb4ab343cec3eb682a9c845c65cb7fbab6"} Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.739901 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"62024578-ee41-4839-9846-9a76d8785d6e","Type":"ContainerStarted","Data":"2cc33d0b7ca3fd352b1fd385a7dc65154e3771208fbe154805b5a64fda9d7541"} Dec 03 14:08:24 crc kubenswrapper[4677]: E1203 14:08:24.741425 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-ovn-sb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="9b7e7e94-5f47-4963-acb0-f40d4ae3ca16" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.741510 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ea48639b-8147-44bb-aa75-96c2ffc0921f","Type":"ContainerStarted","Data":"52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038"} Dec 03 14:08:24 crc kubenswrapper[4677]: E1203 14:08:24.741730 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="62024578-ee41-4839-9846-9a76d8785d6e" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.744387 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"315da922-99be-44c7-81d8-24f0cf55490a","Type":"ContainerStarted","Data":"a0583341d41688452dc42a9a42e55645154003680f58e50ed9e609b979cbf91f"} Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.746339 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.746331 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bb459f7-vkrds" event={"ID":"048d4741-5beb-457e-87b1-e3e215ac9012","Type":"ContainerDied","Data":"f24a546a078832ce93898e1a5ba55109e6405495d6d845e1cc8ccb41a43feb84"} Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.746399 4677 scope.go:117] "RemoveContainer" containerID="97bc3554534b6a204b42c10ef282893412d0228537290d83522bf1bfa1005568" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.786154 4677 scope.go:117] "RemoveContainer" containerID="74e62feded806add1645de56c1938a882edd233eb024bb5e340a73d56680437f" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.791637 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.791671 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048d4741-5beb-457e-87b1-e3e215ac9012-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.791683 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz25z\" (UniqueName: \"kubernetes.io/projected/048d4741-5beb-457e-87b1-e3e215ac9012-kube-api-access-hz25z\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.882777 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778bb459f7-vkrds"] Dec 03 14:08:24 crc kubenswrapper[4677]: I1203 14:08:24.889599 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-778bb459f7-vkrds"] Dec 03 14:08:25 crc kubenswrapper[4677]: I1203 14:08:25.755155 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-msfg2" event={"ID":"d11b3a89-22f0-4768-bc37-701a49737694","Type":"ContainerStarted","Data":"cbd53ce97425c1346afd34bcc4f36a82590748de96752e64114c5d18ccae589d"} Dec 03 14:08:25 crc kubenswrapper[4677]: I1203 14:08:25.755267 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:08:25 crc kubenswrapper[4677]: I1203 14:08:25.755310 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:08:25 crc kubenswrapper[4677]: E1203 14:08:25.758010 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-ovn-sb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="9b7e7e94-5f47-4963-acb0-f40d4ae3ca16" Dec 03 14:08:25 crc kubenswrapper[4677]: E1203 14:08:25.758757 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="62024578-ee41-4839-9846-9a76d8785d6e" Dec 03 14:08:25 crc kubenswrapper[4677]: I1203 14:08:25.808609 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-msfg2" podStartSLOduration=14.877753088 podStartE2EDuration="30.808587914s" podCreationTimestamp="2025-12-03 14:07:55 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.562669001 +0000 UTC m=+1276.309001456" lastFinishedPulling="2025-12-03 14:08:21.493503827 +0000 UTC m=+1292.239836282" observedRunningTime="2025-12-03 14:08:25.788830218 +0000 UTC m=+1296.535162693" watchObservedRunningTime="2025-12-03 14:08:25.808587914 +0000 UTC m=+1296.554920409" Dec 03 14:08:25 crc kubenswrapper[4677]: I1203 14:08:25.988731 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048d4741-5beb-457e-87b1-e3e215ac9012" path="/var/lib/kubelet/pods/048d4741-5beb-457e-87b1-e3e215ac9012/volumes" Dec 03 14:08:31 crc kubenswrapper[4677]: I1203 14:08:31.105167 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 14:08:32 crc kubenswrapper[4677]: I1203 14:08:32.819423 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"072771d7-84c7-4274-8a3d-194f18a09b24","Type":"ContainerStarted","Data":"162aa5ad60a203b96ac043a6185ad7b0e21599983bb3281ccb2f755620547c36"} Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.200234 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-685b674d5c-2jn4s"] Dec 03 14:08:33 crc kubenswrapper[4677]: E1203 14:08:33.201099 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756512c9-e905-4c54-bbc3-73eb9e4b2e45" containerName="init" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.201124 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="756512c9-e905-4c54-bbc3-73eb9e4b2e45" containerName="init" Dec 03 14:08:33 crc kubenswrapper[4677]: E1203 14:08:33.201148 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048d4741-5beb-457e-87b1-e3e215ac9012" containerName="init" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.201156 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="048d4741-5beb-457e-87b1-e3e215ac9012" containerName="init" Dec 03 14:08:33 crc kubenswrapper[4677]: E1203 14:08:33.201181 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048d4741-5beb-457e-87b1-e3e215ac9012" containerName="dnsmasq-dns" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.201189 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="048d4741-5beb-457e-87b1-e3e215ac9012" containerName="dnsmasq-dns" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.201372 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="756512c9-e905-4c54-bbc3-73eb9e4b2e45" containerName="init" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.201386 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="048d4741-5beb-457e-87b1-e3e215ac9012" containerName="dnsmasq-dns" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.202482 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.217108 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-685b674d5c-2jn4s"] Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.350135 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-dns-svc\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.350542 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrzq5\" (UniqueName: \"kubernetes.io/projected/08e651f1-3479-4e30-88d9-6977772e83b5-kube-api-access-xrzq5\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.350583 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-config\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.451939 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrzq5\" (UniqueName: \"kubernetes.io/projected/08e651f1-3479-4e30-88d9-6977772e83b5-kube-api-access-xrzq5\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.452001 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-config\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.452041 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-dns-svc\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.452990 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-dns-svc\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.452999 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-config\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.468527 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrzq5\" (UniqueName: \"kubernetes.io/projected/08e651f1-3479-4e30-88d9-6977772e83b5-kube-api-access-xrzq5\") pod \"dnsmasq-dns-685b674d5c-2jn4s\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.529113 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.836051 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a5ae3526-170f-474a-acce-41e5889470c8","Type":"ContainerStarted","Data":"9ee4d63beb19ba239ec2d7cd55fb3b63c195cb5258ed362eda99af04dbba03fd"} Dec 03 14:08:33 crc kubenswrapper[4677]: I1203 14:08:33.961095 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-685b674d5c-2jn4s"] Dec 03 14:08:33 crc kubenswrapper[4677]: W1203 14:08:33.969422 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08e651f1_3479_4e30_88d9_6977772e83b5.slice/crio-7c429d111be7906c33511a4d0c23b9ac1647deb79c4e50ff8cbb8d6b54237001 WatchSource:0}: Error finding container 7c429d111be7906c33511a4d0c23b9ac1647deb79c4e50ff8cbb8d6b54237001: Status 404 returned error can't find the container with id 7c429d111be7906c33511a4d0c23b9ac1647deb79c4e50ff8cbb8d6b54237001 Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.372883 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.381794 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.383640 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.384224 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.384665 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-89hmm" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.385827 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.412814 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.472912 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b25cf219-67f3-4764-a5da-3dbe89f71641-cache\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.473180 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b25cf219-67f3-4764-a5da-3dbe89f71641-lock\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.473227 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.473315 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rlwl\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-kube-api-access-7rlwl\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.473501 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.574639 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b25cf219-67f3-4764-a5da-3dbe89f71641-cache\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.574724 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b25cf219-67f3-4764-a5da-3dbe89f71641-lock\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.574747 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.574787 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rlwl\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-kube-api-access-7rlwl\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.574868 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.575196 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: E1203 14:08:34.575543 4677 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:08:34 crc kubenswrapper[4677]: E1203 14:08:34.575566 4677 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:08:34 crc kubenswrapper[4677]: E1203 14:08:34.575613 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift podName:b25cf219-67f3-4764-a5da-3dbe89f71641 nodeName:}" failed. No retries permitted until 2025-12-03 14:08:35.075593088 +0000 UTC m=+1305.821925543 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift") pod "swift-storage-0" (UID: "b25cf219-67f3-4764-a5da-3dbe89f71641") : configmap "swift-ring-files" not found Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.575543 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b25cf219-67f3-4764-a5da-3dbe89f71641-cache\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.575848 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b25cf219-67f3-4764-a5da-3dbe89f71641-lock\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.594866 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rlwl\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-kube-api-access-7rlwl\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.600821 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.845355 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8afc071f-9f1a-4830-bd22-a1c291e5ae53","Type":"ContainerStarted","Data":"34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c"} Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.846096 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.847053 4677 generic.go:334] "Generic (PLEG): container finished" podID="08e651f1-3479-4e30-88d9-6977772e83b5" containerID="300ef3bf60bddd05b80ed0ec273ae0286a71f50ecf3636e29dd709d502bee2b7" exitCode=0 Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.847100 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" event={"ID":"08e651f1-3479-4e30-88d9-6977772e83b5","Type":"ContainerDied","Data":"300ef3bf60bddd05b80ed0ec273ae0286a71f50ecf3636e29dd709d502bee2b7"} Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.847140 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" event={"ID":"08e651f1-3479-4e30-88d9-6977772e83b5","Type":"ContainerStarted","Data":"7c429d111be7906c33511a4d0c23b9ac1647deb79c4e50ff8cbb8d6b54237001"} Dec 03 14:08:34 crc kubenswrapper[4677]: I1203 14:08:34.866642 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.988268386 podStartE2EDuration="42.866625782s" podCreationTimestamp="2025-12-03 14:07:52 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.53135753 +0000 UTC m=+1276.277689985" lastFinishedPulling="2025-12-03 14:08:34.409714906 +0000 UTC m=+1305.156047381" observedRunningTime="2025-12-03 14:08:34.861697136 +0000 UTC m=+1305.608029611" watchObservedRunningTime="2025-12-03 14:08:34.866625782 +0000 UTC m=+1305.612958237" Dec 03 14:08:35 crc kubenswrapper[4677]: I1203 14:08:35.082748 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:35 crc kubenswrapper[4677]: E1203 14:08:35.084288 4677 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:08:35 crc kubenswrapper[4677]: E1203 14:08:35.084325 4677 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:08:35 crc kubenswrapper[4677]: E1203 14:08:35.084376 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift podName:b25cf219-67f3-4764-a5da-3dbe89f71641 nodeName:}" failed. No retries permitted until 2025-12-03 14:08:36.084355571 +0000 UTC m=+1306.830688126 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift") pod "swift-storage-0" (UID: "b25cf219-67f3-4764-a5da-3dbe89f71641") : configmap "swift-ring-files" not found Dec 03 14:08:35 crc kubenswrapper[4677]: I1203 14:08:35.859446 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" event={"ID":"08e651f1-3479-4e30-88d9-6977772e83b5","Type":"ContainerStarted","Data":"0790f33959615970b5a242da4ddbeed3e034a92c61a891fe049882b755f34be8"} Dec 03 14:08:36 crc kubenswrapper[4677]: I1203 14:08:36.100812 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:36 crc kubenswrapper[4677]: E1203 14:08:36.101010 4677 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:08:36 crc kubenswrapper[4677]: E1203 14:08:36.101034 4677 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:08:36 crc kubenswrapper[4677]: E1203 14:08:36.101103 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift podName:b25cf219-67f3-4764-a5da-3dbe89f71641 nodeName:}" failed. No retries permitted until 2025-12-03 14:08:38.101076856 +0000 UTC m=+1308.847409311 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift") pod "swift-storage-0" (UID: "b25cf219-67f3-4764-a5da-3dbe89f71641") : configmap "swift-ring-files" not found Dec 03 14:08:36 crc kubenswrapper[4677]: I1203 14:08:36.865942 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:37 crc kubenswrapper[4677]: I1203 14:08:37.881839 4677 generic.go:334] "Generic (PLEG): container finished" podID="072771d7-84c7-4274-8a3d-194f18a09b24" containerID="162aa5ad60a203b96ac043a6185ad7b0e21599983bb3281ccb2f755620547c36" exitCode=0 Dec 03 14:08:37 crc kubenswrapper[4677]: I1203 14:08:37.881887 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"072771d7-84c7-4274-8a3d-194f18a09b24","Type":"ContainerDied","Data":"162aa5ad60a203b96ac043a6185ad7b0e21599983bb3281ccb2f755620547c36"} Dec 03 14:08:37 crc kubenswrapper[4677]: I1203 14:08:37.907206 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" podStartSLOduration=4.907176551 podStartE2EDuration="4.907176551s" podCreationTimestamp="2025-12-03 14:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:35.87523248 +0000 UTC m=+1306.621564935" watchObservedRunningTime="2025-12-03 14:08:37.907176551 +0000 UTC m=+1308.653509046" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.134773 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:38 crc kubenswrapper[4677]: E1203 14:08:38.134971 4677 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:08:38 crc kubenswrapper[4677]: E1203 14:08:38.135252 4677 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:08:38 crc kubenswrapper[4677]: E1203 14:08:38.135318 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift podName:b25cf219-67f3-4764-a5da-3dbe89f71641 nodeName:}" failed. No retries permitted until 2025-12-03 14:08:42.135298586 +0000 UTC m=+1312.881631041 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift") pod "swift-storage-0" (UID: "b25cf219-67f3-4764-a5da-3dbe89f71641") : configmap "swift-ring-files" not found Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.301278 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-qqscx"] Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.305508 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.308224 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.309061 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.309220 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.314482 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qqscx"] Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.439875 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-scripts\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.439960 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-combined-ca-bundle\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.439992 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-dispersionconf\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.440092 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-swiftconf\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.440146 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d39bf749-8b60-4833-97e0-b00791e11467-etc-swift\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.440181 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-ring-data-devices\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.440206 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qlfs\" (UniqueName: \"kubernetes.io/projected/d39bf749-8b60-4833-97e0-b00791e11467-kube-api-access-7qlfs\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.542304 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-scripts\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.542424 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-combined-ca-bundle\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.542490 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-dispersionconf\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.542665 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-swiftconf\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.542778 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d39bf749-8b60-4833-97e0-b00791e11467-etc-swift\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.542899 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-ring-data-devices\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.543025 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qlfs\" (UniqueName: \"kubernetes.io/projected/d39bf749-8b60-4833-97e0-b00791e11467-kube-api-access-7qlfs\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.543297 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d39bf749-8b60-4833-97e0-b00791e11467-etc-swift\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.543366 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-scripts\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.543853 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-ring-data-devices\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.546719 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-dispersionconf\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.547822 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-combined-ca-bundle\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.553413 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-swiftconf\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.561801 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qlfs\" (UniqueName: \"kubernetes.io/projected/d39bf749-8b60-4833-97e0-b00791e11467-kube-api-access-7qlfs\") pod \"swift-ring-rebalance-qqscx\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.644382 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.901135 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"072771d7-84c7-4274-8a3d-194f18a09b24","Type":"ContainerStarted","Data":"ba610ba4ec3b78272d211b865f009f23347fa56dca659762ae084a85e6effcd9"} Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.902676 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerStarted","Data":"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb"} Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.905086 4677 generic.go:334] "Generic (PLEG): container finished" podID="a5ae3526-170f-474a-acce-41e5889470c8" containerID="9ee4d63beb19ba239ec2d7cd55fb3b63c195cb5258ed362eda99af04dbba03fd" exitCode=0 Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.905152 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a5ae3526-170f-474a-acce-41e5889470c8","Type":"ContainerDied","Data":"9ee4d63beb19ba239ec2d7cd55fb3b63c195cb5258ed362eda99af04dbba03fd"} Dec 03 14:08:38 crc kubenswrapper[4677]: I1203 14:08:38.926070 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=23.19404627 podStartE2EDuration="49.926053782s" podCreationTimestamp="2025-12-03 14:07:49 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.364815395 +0000 UTC m=+1276.111147840" lastFinishedPulling="2025-12-03 14:08:32.096822897 +0000 UTC m=+1302.843155352" observedRunningTime="2025-12-03 14:08:38.92052541 +0000 UTC m=+1309.666857875" watchObservedRunningTime="2025-12-03 14:08:38.926053782 +0000 UTC m=+1309.672386237" Dec 03 14:08:39 crc kubenswrapper[4677]: I1203 14:08:39.070786 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qqscx"] Dec 03 14:08:39 crc kubenswrapper[4677]: W1203 14:08:39.076790 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd39bf749_8b60_4833_97e0_b00791e11467.slice/crio-193a2684c99c670c8bd63dcc776127459cafe6868e0e7e95ca4359919b806221 WatchSource:0}: Error finding container 193a2684c99c670c8bd63dcc776127459cafe6868e0e7e95ca4359919b806221: Status 404 returned error can't find the container with id 193a2684c99c670c8bd63dcc776127459cafe6868e0e7e95ca4359919b806221 Dec 03 14:08:39 crc kubenswrapper[4677]: I1203 14:08:39.917540 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a5ae3526-170f-474a-acce-41e5889470c8","Type":"ContainerStarted","Data":"5957dc7e31f96ef84f9a5abfa99788d724656e85a72d8c83e37babd072c809f9"} Dec 03 14:08:39 crc kubenswrapper[4677]: I1203 14:08:39.920812 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qqscx" event={"ID":"d39bf749-8b60-4833-97e0-b00791e11467","Type":"ContainerStarted","Data":"193a2684c99c670c8bd63dcc776127459cafe6868e0e7e95ca4359919b806221"} Dec 03 14:08:39 crc kubenswrapper[4677]: I1203 14:08:39.930445 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9b7e7e94-5f47-4963-acb0-f40d4ae3ca16","Type":"ContainerStarted","Data":"467bea8c4f6a6bd70ffc5a763fe58a912cce0eea3c2b28294ffed00e52aa32ea"} Dec 03 14:08:39 crc kubenswrapper[4677]: I1203 14:08:39.942107 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371984.912687 podStartE2EDuration="51.94208759s" podCreationTimestamp="2025-12-03 14:07:48 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.192136107 +0000 UTC m=+1275.938468562" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:39.939011722 +0000 UTC m=+1310.685344207" watchObservedRunningTime="2025-12-03 14:08:39.94208759 +0000 UTC m=+1310.688420045" Dec 03 14:08:39 crc kubenswrapper[4677]: I1203 14:08:39.965504 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.375091546 podStartE2EDuration="40.965484828s" podCreationTimestamp="2025-12-03 14:07:59 +0000 UTC" firstStartedPulling="2025-12-03 14:08:07.480201937 +0000 UTC m=+1278.226534392" lastFinishedPulling="2025-12-03 14:08:39.070595219 +0000 UTC m=+1309.816927674" observedRunningTime="2025-12-03 14:08:39.96281859 +0000 UTC m=+1310.709151055" watchObservedRunningTime="2025-12-03 14:08:39.965484828 +0000 UTC m=+1310.711817283" Dec 03 14:08:40 crc kubenswrapper[4677]: I1203 14:08:40.590810 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:40 crc kubenswrapper[4677]: I1203 14:08:40.788579 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 14:08:40 crc kubenswrapper[4677]: I1203 14:08:40.788642 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.219719 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:42 crc kubenswrapper[4677]: E1203 14:08:42.220017 4677 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:08:42 crc kubenswrapper[4677]: E1203 14:08:42.220256 4677 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:08:42 crc kubenswrapper[4677]: E1203 14:08:42.220398 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift podName:b25cf219-67f3-4764-a5da-3dbe89f71641 nodeName:}" failed. No retries permitted until 2025-12-03 14:08:50.220307022 +0000 UTC m=+1320.966639487 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift") pod "swift-storage-0" (UID: "b25cf219-67f3-4764-a5da-3dbe89f71641") : configmap "swift-ring-files" not found Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.423397 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-bfpgf"] Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.425233 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.428599 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.433644 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bfpgf"] Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.525026 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08873f8c-1b49-42cf-914c-150fd7ec7b09-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.525087 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08873f8c-1b49-42cf-914c-150fd7ec7b09-combined-ca-bundle\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.525193 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/08873f8c-1b49-42cf-914c-150fd7ec7b09-ovs-rundir\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.525229 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/08873f8c-1b49-42cf-914c-150fd7ec7b09-ovn-rundir\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.525261 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r5sj\" (UniqueName: \"kubernetes.io/projected/08873f8c-1b49-42cf-914c-150fd7ec7b09-kube-api-access-9r5sj\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.525294 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08873f8c-1b49-42cf-914c-150fd7ec7b09-config\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.582450 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-685b674d5c-2jn4s"] Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.584121 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.582706 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" containerName="dnsmasq-dns" containerID="cri-o://0790f33959615970b5a242da4ddbeed3e034a92c61a891fe049882b755f34be8" gracePeriod=10 Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.590134 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.626778 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08873f8c-1b49-42cf-914c-150fd7ec7b09-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.626834 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08873f8c-1b49-42cf-914c-150fd7ec7b09-combined-ca-bundle\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.626921 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/08873f8c-1b49-42cf-914c-150fd7ec7b09-ovs-rundir\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.626973 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/08873f8c-1b49-42cf-914c-150fd7ec7b09-ovn-rundir\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.627003 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r5sj\" (UniqueName: \"kubernetes.io/projected/08873f8c-1b49-42cf-914c-150fd7ec7b09-kube-api-access-9r5sj\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.627034 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08873f8c-1b49-42cf-914c-150fd7ec7b09-config\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.628040 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/08873f8c-1b49-42cf-914c-150fd7ec7b09-ovs-rundir\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.628059 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/08873f8c-1b49-42cf-914c-150fd7ec7b09-ovn-rundir\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.629205 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08873f8c-1b49-42cf-914c-150fd7ec7b09-config\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.639675 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-765596c6bf-4tsph"] Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.641145 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.648264 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08873f8c-1b49-42cf-914c-150fd7ec7b09-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.651744 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08873f8c-1b49-42cf-914c-150fd7ec7b09-combined-ca-bundle\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.661218 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.663809 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-765596c6bf-4tsph"] Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.675424 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r5sj\" (UniqueName: \"kubernetes.io/projected/08873f8c-1b49-42cf-914c-150fd7ec7b09-kube-api-access-9r5sj\") pod \"ovn-controller-metrics-bfpgf\" (UID: \"08873f8c-1b49-42cf-914c-150fd7ec7b09\") " pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.677111 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.774497 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bfpgf" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.830274 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lj6v\" (UniqueName: \"kubernetes.io/projected/7e42da9b-0f14-4c95-be7d-1cd047c8650b-kube-api-access-7lj6v\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.830606 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-dns-svc\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.830640 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-config\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.830688 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-ovsdbserver-nb\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.856764 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-765596c6bf-4tsph"] Dec 03 14:08:42 crc kubenswrapper[4677]: E1203 14:08:42.857341 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-7lj6v ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-765596c6bf-4tsph" podUID="7e42da9b-0f14-4c95-be7d-1cd047c8650b" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.885085 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c98758bb9-jwrht"] Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.886817 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.894312 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c98758bb9-jwrht"] Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.895480 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.932007 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lj6v\" (UniqueName: \"kubernetes.io/projected/7e42da9b-0f14-4c95-be7d-1cd047c8650b-kube-api-access-7lj6v\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.932081 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-dns-svc\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.932128 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-config\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.932181 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-ovsdbserver-nb\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.933236 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-dns-svc\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.933283 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-ovsdbserver-nb\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.933353 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-config\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.950382 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lj6v\" (UniqueName: \"kubernetes.io/projected/7e42da9b-0f14-4c95-be7d-1cd047c8650b-kube-api-access-7lj6v\") pod \"dnsmasq-dns-765596c6bf-4tsph\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.960298 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:42 crc kubenswrapper[4677]: I1203 14:08:42.990354 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.011397 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.033940 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkt8t\" (UniqueName: \"kubernetes.io/projected/e0ebab4b-62db-4b77-b730-0413b51bdeaf-kube-api-access-fkt8t\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.034003 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-dns-svc\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.034023 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-nb\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.034049 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-config\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.034414 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-sb\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136012 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lj6v\" (UniqueName: \"kubernetes.io/projected/7e42da9b-0f14-4c95-be7d-1cd047c8650b-kube-api-access-7lj6v\") pod \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136067 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-dns-svc\") pod \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136135 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-config\") pod \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136215 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-ovsdbserver-nb\") pod \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\" (UID: \"7e42da9b-0f14-4c95-be7d-1cd047c8650b\") " Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136501 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-sb\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136574 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e42da9b-0f14-4c95-be7d-1cd047c8650b" (UID: "7e42da9b-0f14-4c95-be7d-1cd047c8650b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136589 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkt8t\" (UniqueName: \"kubernetes.io/projected/e0ebab4b-62db-4b77-b730-0413b51bdeaf-kube-api-access-fkt8t\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136637 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-dns-svc\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136656 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-nb\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136677 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-config\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136672 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e42da9b-0f14-4c95-be7d-1cd047c8650b" (UID: "7e42da9b-0f14-4c95-be7d-1cd047c8650b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136828 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.136842 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.137278 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-config" (OuterVolumeSpecName: "config") pod "7e42da9b-0f14-4c95-be7d-1cd047c8650b" (UID: "7e42da9b-0f14-4c95-be7d-1cd047c8650b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.137566 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-sb\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.137777 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-dns-svc\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.137888 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-config\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.138114 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-nb\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.138984 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e42da9b-0f14-4c95-be7d-1cd047c8650b-kube-api-access-7lj6v" (OuterVolumeSpecName: "kube-api-access-7lj6v") pod "7e42da9b-0f14-4c95-be7d-1cd047c8650b" (UID: "7e42da9b-0f14-4c95-be7d-1cd047c8650b"). InnerVolumeSpecName "kube-api-access-7lj6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.151514 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkt8t\" (UniqueName: \"kubernetes.io/projected/e0ebab4b-62db-4b77-b730-0413b51bdeaf-kube-api-access-fkt8t\") pod \"dnsmasq-dns-5c98758bb9-jwrht\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.203397 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.238149 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lj6v\" (UniqueName: \"kubernetes.io/projected/7e42da9b-0f14-4c95-be7d-1cd047c8650b-kube-api-access-7lj6v\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.238182 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e42da9b-0f14-4c95-be7d-1cd047c8650b-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.530234 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Dec 03 14:08:43 crc kubenswrapper[4677]: I1203 14:08:43.967570 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-765596c6bf-4tsph" Dec 03 14:08:44 crc kubenswrapper[4677]: I1203 14:08:44.011914 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-765596c6bf-4tsph"] Dec 03 14:08:44 crc kubenswrapper[4677]: I1203 14:08:44.021894 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-765596c6bf-4tsph"] Dec 03 14:08:44 crc kubenswrapper[4677]: I1203 14:08:44.981141 4677 generic.go:334] "Generic (PLEG): container finished" podID="08e651f1-3479-4e30-88d9-6977772e83b5" containerID="0790f33959615970b5a242da4ddbeed3e034a92c61a891fe049882b755f34be8" exitCode=0 Dec 03 14:08:44 crc kubenswrapper[4677]: I1203 14:08:44.981319 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" event={"ID":"08e651f1-3479-4e30-88d9-6977772e83b5","Type":"ContainerDied","Data":"0790f33959615970b5a242da4ddbeed3e034a92c61a891fe049882b755f34be8"} Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.268925 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.370562 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-dns-svc\") pod \"08e651f1-3479-4e30-88d9-6977772e83b5\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.370597 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-config\") pod \"08e651f1-3479-4e30-88d9-6977772e83b5\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.370701 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrzq5\" (UniqueName: \"kubernetes.io/projected/08e651f1-3479-4e30-88d9-6977772e83b5-kube-api-access-xrzq5\") pod \"08e651f1-3479-4e30-88d9-6977772e83b5\" (UID: \"08e651f1-3479-4e30-88d9-6977772e83b5\") " Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.377070 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08e651f1-3479-4e30-88d9-6977772e83b5-kube-api-access-xrzq5" (OuterVolumeSpecName: "kube-api-access-xrzq5") pod "08e651f1-3479-4e30-88d9-6977772e83b5" (UID: "08e651f1-3479-4e30-88d9-6977772e83b5"). InnerVolumeSpecName "kube-api-access-xrzq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.413726 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08e651f1-3479-4e30-88d9-6977772e83b5" (UID: "08e651f1-3479-4e30-88d9-6977772e83b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.414867 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bfpgf"] Dec 03 14:08:45 crc kubenswrapper[4677]: W1203 14:08:45.418028 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08873f8c_1b49_42cf_914c_150fd7ec7b09.slice/crio-46808b6871f0efe7e6c75129ee931d4f611a0227b70ebefd558a61202251f111 WatchSource:0}: Error finding container 46808b6871f0efe7e6c75129ee931d4f611a0227b70ebefd558a61202251f111: Status 404 returned error can't find the container with id 46808b6871f0efe7e6c75129ee931d4f611a0227b70ebefd558a61202251f111 Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.423193 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-config" (OuterVolumeSpecName: "config") pod "08e651f1-3479-4e30-88d9-6977772e83b5" (UID: "08e651f1-3479-4e30-88d9-6977772e83b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.472470 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.472500 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e651f1-3479-4e30-88d9-6977772e83b5-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.472513 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrzq5\" (UniqueName: \"kubernetes.io/projected/08e651f1-3479-4e30-88d9-6977772e83b5-kube-api-access-xrzq5\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.496803 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c98758bb9-jwrht"] Dec 03 14:08:45 crc kubenswrapper[4677]: W1203 14:08:45.507778 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0ebab4b_62db_4b77_b730_0413b51bdeaf.slice/crio-8fcfa123058cc29ae9f50e89a458a6e6809ba67a34f713324bd10fe0bf31b5b2 WatchSource:0}: Error finding container 8fcfa123058cc29ae9f50e89a458a6e6809ba67a34f713324bd10fe0bf31b5b2: Status 404 returned error can't find the container with id 8fcfa123058cc29ae9f50e89a458a6e6809ba67a34f713324bd10fe0bf31b5b2 Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.638264 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.986208 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e42da9b-0f14-4c95-be7d-1cd047c8650b" path="/var/lib/kubelet/pods/7e42da9b-0f14-4c95-be7d-1cd047c8650b/volumes" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.990800 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.990800 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685b674d5c-2jn4s" event={"ID":"08e651f1-3479-4e30-88d9-6977772e83b5","Type":"ContainerDied","Data":"7c429d111be7906c33511a4d0c23b9ac1647deb79c4e50ff8cbb8d6b54237001"} Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.991011 4677 scope.go:117] "RemoveContainer" containerID="0790f33959615970b5a242da4ddbeed3e034a92c61a891fe049882b755f34be8" Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.994415 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bfpgf" event={"ID":"08873f8c-1b49-42cf-914c-150fd7ec7b09","Type":"ContainerStarted","Data":"7e528d95d6bc78dcbdd0d3772514cf2f90c2e69ef6797ac162bd3ae553667870"} Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.994459 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bfpgf" event={"ID":"08873f8c-1b49-42cf-914c-150fd7ec7b09","Type":"ContainerStarted","Data":"46808b6871f0efe7e6c75129ee931d4f611a0227b70ebefd558a61202251f111"} Dec 03 14:08:45 crc kubenswrapper[4677]: I1203 14:08:45.997425 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qqscx" event={"ID":"d39bf749-8b60-4833-97e0-b00791e11467","Type":"ContainerStarted","Data":"1986017e383b5d56c89a52892ee9f9ae17e8953c16a04f1d673fdff09b140110"} Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.010250 4677 scope.go:117] "RemoveContainer" containerID="300ef3bf60bddd05b80ed0ec273ae0286a71f50ecf3636e29dd709d502bee2b7" Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.016345 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"62024578-ee41-4839-9846-9a76d8785d6e","Type":"ContainerStarted","Data":"19fabb9664a7b8536d8acffcba698aa13a4293ff6a896a4613790c2bad3afcc5"} Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.018841 4677 generic.go:334] "Generic (PLEG): container finished" podID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerID="ad1edec02b060afc9ba51b3a8a2d1a56e70d94e171d3c3dc8307896909ffc7f5" exitCode=0 Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.019124 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" event={"ID":"e0ebab4b-62db-4b77-b730-0413b51bdeaf","Type":"ContainerDied","Data":"ad1edec02b060afc9ba51b3a8a2d1a56e70d94e171d3c3dc8307896909ffc7f5"} Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.019223 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" event={"ID":"e0ebab4b-62db-4b77-b730-0413b51bdeaf","Type":"ContainerStarted","Data":"8fcfa123058cc29ae9f50e89a458a6e6809ba67a34f713324bd10fe0bf31b5b2"} Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.028010 4677 generic.go:334] "Generic (PLEG): container finished" podID="db9e192a-61af-41fa-882c-81879f2a556a" containerID="dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb" exitCode=0 Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.028053 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerDied","Data":"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb"} Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.036284 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-bfpgf" podStartSLOduration=4.036269515 podStartE2EDuration="4.036269515s" podCreationTimestamp="2025-12-03 14:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:46.032313724 +0000 UTC m=+1316.778646179" watchObservedRunningTime="2025-12-03 14:08:46.036269515 +0000 UTC m=+1316.782601970" Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.169357 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-qqscx" podStartSLOduration=2.23179512 podStartE2EDuration="8.169335919s" podCreationTimestamp="2025-12-03 14:08:38 +0000 UTC" firstStartedPulling="2025-12-03 14:08:39.079443215 +0000 UTC m=+1309.825775670" lastFinishedPulling="2025-12-03 14:08:45.016984014 +0000 UTC m=+1315.763316469" observedRunningTime="2025-12-03 14:08:46.133399379 +0000 UTC m=+1316.879731844" watchObservedRunningTime="2025-12-03 14:08:46.169335919 +0000 UTC m=+1316.915668364" Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.209133 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-685b674d5c-2jn4s"] Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.215464 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-685b674d5c-2jn4s"] Dec 03 14:08:46 crc kubenswrapper[4677]: I1203 14:08:46.216629 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=13.690047852 podStartE2EDuration="51.216607308s" podCreationTimestamp="2025-12-03 14:07:55 +0000 UTC" firstStartedPulling="2025-12-03 14:08:07.480111905 +0000 UTC m=+1278.226444360" lastFinishedPulling="2025-12-03 14:08:45.006671331 +0000 UTC m=+1315.753003816" observedRunningTime="2025-12-03 14:08:46.183671325 +0000 UTC m=+1316.930003810" watchObservedRunningTime="2025-12-03 14:08:46.216607308 +0000 UTC m=+1316.962939763" Dec 03 14:08:47 crc kubenswrapper[4677]: I1203 14:08:47.036402 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" event={"ID":"e0ebab4b-62db-4b77-b730-0413b51bdeaf","Type":"ContainerStarted","Data":"1d923eb357efa003508d4e717adee4aa5a17e8572cf26b077ee79b1c77ca8dd8"} Dec 03 14:08:47 crc kubenswrapper[4677]: I1203 14:08:47.036731 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:47 crc kubenswrapper[4677]: I1203 14:08:47.065317 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" podStartSLOduration=5.065299084 podStartE2EDuration="5.065299084s" podCreationTimestamp="2025-12-03 14:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:47.062611477 +0000 UTC m=+1317.808943942" watchObservedRunningTime="2025-12-03 14:08:47.065299084 +0000 UTC m=+1317.811631539" Dec 03 14:08:47 crc kubenswrapper[4677]: I1203 14:08:47.149277 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 14:08:47 crc kubenswrapper[4677]: I1203 14:08:47.464466 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 14:08:47 crc kubenswrapper[4677]: I1203 14:08:47.602627 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 14:08:47 crc kubenswrapper[4677]: I1203 14:08:47.987244 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" path="/var/lib/kubelet/pods/08e651f1-3479-4e30-88d9-6977772e83b5/volumes" Dec 03 14:08:48 crc kubenswrapper[4677]: I1203 14:08:48.150268 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 14:08:48 crc kubenswrapper[4677]: I1203 14:08:48.202868 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 14:08:49 crc kubenswrapper[4677]: I1203 14:08:49.447085 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 14:08:49 crc kubenswrapper[4677]: I1203 14:08:49.447148 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 14:08:49 crc kubenswrapper[4677]: I1203 14:08:49.703705 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.321187 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:08:50 crc kubenswrapper[4677]: E1203 14:08:50.321352 4677 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:08:50 crc kubenswrapper[4677]: E1203 14:08:50.321452 4677 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:08:50 crc kubenswrapper[4677]: E1203 14:08:50.321507 4677 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift podName:b25cf219-67f3-4764-a5da-3dbe89f71641 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:06.32148937 +0000 UTC m=+1337.067821815 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift") pod "swift-storage-0" (UID: "b25cf219-67f3-4764-a5da-3dbe89f71641") : configmap "swift-ring-files" not found Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.368426 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.744123 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bbef-account-create-update-8btps"] Dec 03 14:08:50 crc kubenswrapper[4677]: E1203 14:08:50.744648 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" containerName="dnsmasq-dns" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.744661 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" containerName="dnsmasq-dns" Dec 03 14:08:50 crc kubenswrapper[4677]: E1203 14:08:50.744690 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" containerName="init" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.744696 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" containerName="init" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.744852 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="08e651f1-3479-4e30-88d9-6977772e83b5" containerName="dnsmasq-dns" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.745416 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.751852 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.761592 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bbef-account-create-update-8btps"] Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.812975 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ztpkm"] Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.814042 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.830317 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-operator-scripts\") pod \"keystone-bbef-account-create-update-8btps\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.830754 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z66sm\" (UniqueName: \"kubernetes.io/projected/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-kube-api-access-z66sm\") pod \"keystone-bbef-account-create-update-8btps\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.836611 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ztpkm"] Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.932396 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldjjv\" (UniqueName: \"kubernetes.io/projected/2994ba2a-3d45-4303-b0c1-9800979e1cb5-kube-api-access-ldjjv\") pod \"keystone-db-create-ztpkm\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.932475 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2994ba2a-3d45-4303-b0c1-9800979e1cb5-operator-scripts\") pod \"keystone-db-create-ztpkm\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.932543 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z66sm\" (UniqueName: \"kubernetes.io/projected/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-kube-api-access-z66sm\") pod \"keystone-bbef-account-create-update-8btps\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.932815 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-operator-scripts\") pod \"keystone-bbef-account-create-update-8btps\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.933807 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-operator-scripts\") pod \"keystone-bbef-account-create-update-8btps\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:50 crc kubenswrapper[4677]: I1203 14:08:50.952522 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z66sm\" (UniqueName: \"kubernetes.io/projected/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-kube-api-access-z66sm\") pod \"keystone-bbef-account-create-update-8btps\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.034257 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2994ba2a-3d45-4303-b0c1-9800979e1cb5-operator-scripts\") pod \"keystone-db-create-ztpkm\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.034517 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldjjv\" (UniqueName: \"kubernetes.io/projected/2994ba2a-3d45-4303-b0c1-9800979e1cb5-kube-api-access-ldjjv\") pod \"keystone-db-create-ztpkm\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.035195 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2994ba2a-3d45-4303-b0c1-9800979e1cb5-operator-scripts\") pod \"keystone-db-create-ztpkm\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.052873 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-zk8qb"] Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.053904 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.063518 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-da07-account-create-update-fpsng"] Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.064631 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.065308 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.071904 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-zk8qb"] Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.073551 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.080393 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldjjv\" (UniqueName: \"kubernetes.io/projected/2994ba2a-3d45-4303-b0c1-9800979e1cb5-kube-api-access-ldjjv\") pod \"keystone-db-create-ztpkm\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.132824 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-da07-account-create-update-fpsng"] Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.133413 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.141397 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79h9f\" (UniqueName: \"kubernetes.io/projected/445b8de4-6203-494a-9448-86f12765390e-kube-api-access-79h9f\") pod \"placement-da07-account-create-update-fpsng\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.141457 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/445b8de4-6203-494a-9448-86f12765390e-operator-scripts\") pod \"placement-da07-account-create-update-fpsng\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.141514 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2bvv\" (UniqueName: \"kubernetes.io/projected/e7c96896-06fc-4875-a6c0-f0f3cec43b90-kube-api-access-k2bvv\") pod \"placement-db-create-zk8qb\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.141793 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c96896-06fc-4875-a6c0-f0f3cec43b90-operator-scripts\") pod \"placement-db-create-zk8qb\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.244099 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c96896-06fc-4875-a6c0-f0f3cec43b90-operator-scripts\") pod \"placement-db-create-zk8qb\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.244521 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79h9f\" (UniqueName: \"kubernetes.io/projected/445b8de4-6203-494a-9448-86f12765390e-kube-api-access-79h9f\") pod \"placement-da07-account-create-update-fpsng\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.244643 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/445b8de4-6203-494a-9448-86f12765390e-operator-scripts\") pod \"placement-da07-account-create-update-fpsng\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.244677 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2bvv\" (UniqueName: \"kubernetes.io/projected/e7c96896-06fc-4875-a6c0-f0f3cec43b90-kube-api-access-k2bvv\") pod \"placement-db-create-zk8qb\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.245328 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c96896-06fc-4875-a6c0-f0f3cec43b90-operator-scripts\") pod \"placement-db-create-zk8qb\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.247852 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/445b8de4-6203-494a-9448-86f12765390e-operator-scripts\") pod \"placement-da07-account-create-update-fpsng\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.270458 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2bvv\" (UniqueName: \"kubernetes.io/projected/e7c96896-06fc-4875-a6c0-f0f3cec43b90-kube-api-access-k2bvv\") pod \"placement-db-create-zk8qb\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.277606 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79h9f\" (UniqueName: \"kubernetes.io/projected/445b8de4-6203-494a-9448-86f12765390e-kube-api-access-79h9f\") pod \"placement-da07-account-create-update-fpsng\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.377736 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.456542 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.626006 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bbef-account-create-update-8btps"] Dec 03 14:08:51 crc kubenswrapper[4677]: W1203 14:08:51.731555 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2994ba2a_3d45_4303_b0c1_9800979e1cb5.slice/crio-1fd1a41dfafea211882e64c49588813a4d53544a0545971d7db425cef619efb6 WatchSource:0}: Error finding container 1fd1a41dfafea211882e64c49588813a4d53544a0545971d7db425cef619efb6: Status 404 returned error can't find the container with id 1fd1a41dfafea211882e64c49588813a4d53544a0545971d7db425cef619efb6 Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.736115 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ztpkm"] Dec 03 14:08:51 crc kubenswrapper[4677]: I1203 14:08:51.960769 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-zk8qb"] Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.086761 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-zk8qb" event={"ID":"e7c96896-06fc-4875-a6c0-f0f3cec43b90","Type":"ContainerStarted","Data":"ea0f3b1f03f89c71d5b9c517dbd21a4744ad6d91b990402266437c755ae39e7b"} Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.087574 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bbef-account-create-update-8btps" event={"ID":"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0","Type":"ContainerStarted","Data":"f6b4aa7bdff2dd6fc3c6f491c6753be7afd2c4f13c9c84af6125ed22984808ef"} Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.088472 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ztpkm" event={"ID":"2994ba2a-3d45-4303-b0c1-9800979e1cb5","Type":"ContainerStarted","Data":"1fd1a41dfafea211882e64c49588813a4d53544a0545971d7db425cef619efb6"} Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.133939 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-da07-account-create-update-fpsng"] Dec 03 14:08:52 crc kubenswrapper[4677]: W1203 14:08:52.146258 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod445b8de4_6203_494a_9448_86f12765390e.slice/crio-ed43f5f6f167d41458f1b25924c065581f3dad1a4d91912842ba83f2f2cbabae WatchSource:0}: Error finding container ed43f5f6f167d41458f1b25924c065581f3dad1a4d91912842ba83f2f2cbabae: Status 404 returned error can't find the container with id ed43f5f6f167d41458f1b25924c065581f3dad1a4d91912842ba83f2f2cbabae Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.202573 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.366967 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.369534 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.372780 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.372964 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.372984 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.373393 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-gqrsc" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.443453 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.471856 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a232e198-17b4-4617-a664-b8db60b0d3fe-scripts\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.471937 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncrtp\" (UniqueName: \"kubernetes.io/projected/a232e198-17b4-4617-a664-b8db60b0d3fe-kube-api-access-ncrtp\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.471979 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.472012 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.472044 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a232e198-17b4-4617-a664-b8db60b0d3fe-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.472091 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.472129 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a232e198-17b4-4617-a664-b8db60b0d3fe-config\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.573795 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.573893 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a232e198-17b4-4617-a664-b8db60b0d3fe-config\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.573941 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a232e198-17b4-4617-a664-b8db60b0d3fe-scripts\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.574046 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncrtp\" (UniqueName: \"kubernetes.io/projected/a232e198-17b4-4617-a664-b8db60b0d3fe-kube-api-access-ncrtp\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.574077 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.574119 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.574164 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a232e198-17b4-4617-a664-b8db60b0d3fe-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.574767 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a232e198-17b4-4617-a664-b8db60b0d3fe-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.575127 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a232e198-17b4-4617-a664-b8db60b0d3fe-config\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.575169 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a232e198-17b4-4617-a664-b8db60b0d3fe-scripts\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.580820 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.581506 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.582870 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a232e198-17b4-4617-a664-b8db60b0d3fe-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.597178 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncrtp\" (UniqueName: \"kubernetes.io/projected/a232e198-17b4-4617-a664-b8db60b0d3fe-kube-api-access-ncrtp\") pod \"ovn-northd-0\" (UID: \"a232e198-17b4-4617-a664-b8db60b0d3fe\") " pod="openstack/ovn-northd-0" Dec 03 14:08:52 crc kubenswrapper[4677]: I1203 14:08:52.709104 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.030074 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-rh642"] Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.031114 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.054423 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-6844-account-create-update-t6r4j"] Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.056230 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.060704 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.085763 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-rh642"] Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.113015 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-6844-account-create-update-t6r4j"] Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.185337 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrbm\" (UniqueName: \"kubernetes.io/projected/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-kube-api-access-zxrbm\") pod \"watcher-db-create-rh642\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.185390 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtfqh\" (UniqueName: \"kubernetes.io/projected/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-kube-api-access-jtfqh\") pod \"watcher-6844-account-create-update-t6r4j\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.185433 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-operator-scripts\") pod \"watcher-db-create-rh642\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.185607 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-operator-scripts\") pod \"watcher-6844-account-create-update-t6r4j\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.188835 4677 generic.go:334] "Generic (PLEG): container finished" podID="e7c96896-06fc-4875-a6c0-f0f3cec43b90" containerID="6799fc13216f7781718fd67889d0ea610dc281c4276d9b182d12d7c03c456825" exitCode=0 Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.189239 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-zk8qb" event={"ID":"e7c96896-06fc-4875-a6c0-f0f3cec43b90","Type":"ContainerDied","Data":"6799fc13216f7781718fd67889d0ea610dc281c4276d9b182d12d7c03c456825"} Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.189674 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.196609 4677 generic.go:334] "Generic (PLEG): container finished" podID="445b8de4-6203-494a-9448-86f12765390e" containerID="4d6bfbd19412d37509ea4d48bf3b92a4f7ef6bdbf2a405ea46c9ca9b55f2aa48" exitCode=0 Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.196664 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-da07-account-create-update-fpsng" event={"ID":"445b8de4-6203-494a-9448-86f12765390e","Type":"ContainerDied","Data":"4d6bfbd19412d37509ea4d48bf3b92a4f7ef6bdbf2a405ea46c9ca9b55f2aa48"} Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.196685 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-da07-account-create-update-fpsng" event={"ID":"445b8de4-6203-494a-9448-86f12765390e","Type":"ContainerStarted","Data":"ed43f5f6f167d41458f1b25924c065581f3dad1a4d91912842ba83f2f2cbabae"} Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.202051 4677 generic.go:334] "Generic (PLEG): container finished" podID="5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0" containerID="360d3c2f92c2bba064ddada31151b6307864efb3175f5e930fe2206dd1b9918a" exitCode=0 Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.202112 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bbef-account-create-update-8btps" event={"ID":"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0","Type":"ContainerDied","Data":"360d3c2f92c2bba064ddada31151b6307864efb3175f5e930fe2206dd1b9918a"} Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.206143 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.206493 4677 generic.go:334] "Generic (PLEG): container finished" podID="2994ba2a-3d45-4303-b0c1-9800979e1cb5" containerID="07bc48d960c258535c88f43adf264d48d0a2a004d9d835a7e8edabbd19a33f8f" exitCode=0 Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.206618 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ztpkm" event={"ID":"2994ba2a-3d45-4303-b0c1-9800979e1cb5","Type":"ContainerDied","Data":"07bc48d960c258535c88f43adf264d48d0a2a004d9d835a7e8edabbd19a33f8f"} Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.287157 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-operator-scripts\") pod \"watcher-6844-account-create-update-t6r4j\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.287228 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxrbm\" (UniqueName: \"kubernetes.io/projected/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-kube-api-access-zxrbm\") pod \"watcher-db-create-rh642\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.287252 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtfqh\" (UniqueName: \"kubernetes.io/projected/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-kube-api-access-jtfqh\") pod \"watcher-6844-account-create-update-t6r4j\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.287293 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-operator-scripts\") pod \"watcher-db-create-rh642\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.288659 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-operator-scripts\") pod \"watcher-6844-account-create-update-t6r4j\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.289439 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-operator-scripts\") pod \"watcher-db-create-rh642\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.316772 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxrbm\" (UniqueName: \"kubernetes.io/projected/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-kube-api-access-zxrbm\") pod \"watcher-db-create-rh642\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.353035 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtfqh\" (UniqueName: \"kubernetes.io/projected/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-kube-api-access-jtfqh\") pod \"watcher-6844-account-create-update-t6r4j\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.376531 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85d946848f-qz6ws"] Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.376791 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" podUID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerName="dnsmasq-dns" containerID="cri-o://0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2" gracePeriod=10 Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.398367 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-rh642" Dec 03 14:08:53 crc kubenswrapper[4677]: I1203 14:08:53.435369 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.020807 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-rh642"] Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.044993 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.109376 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-dns-svc\") pod \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.109448 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf2ct\" (UniqueName: \"kubernetes.io/projected/70265ca5-cffd-42a9-b519-5c5eed5fc29a-kube-api-access-wf2ct\") pod \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.110269 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-config\") pod \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\" (UID: \"70265ca5-cffd-42a9-b519-5c5eed5fc29a\") " Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.118374 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70265ca5-cffd-42a9-b519-5c5eed5fc29a-kube-api-access-wf2ct" (OuterVolumeSpecName: "kube-api-access-wf2ct") pod "70265ca5-cffd-42a9-b519-5c5eed5fc29a" (UID: "70265ca5-cffd-42a9-b519-5c5eed5fc29a"). InnerVolumeSpecName "kube-api-access-wf2ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.137753 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-6844-account-create-update-t6r4j"] Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.156859 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "70265ca5-cffd-42a9-b519-5c5eed5fc29a" (UID: "70265ca5-cffd-42a9-b519-5c5eed5fc29a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.212050 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.212089 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf2ct\" (UniqueName: \"kubernetes.io/projected/70265ca5-cffd-42a9-b519-5c5eed5fc29a-kube-api-access-wf2ct\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.213527 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-config" (OuterVolumeSpecName: "config") pod "70265ca5-cffd-42a9-b519-5c5eed5fc29a" (UID: "70265ca5-cffd-42a9-b519-5c5eed5fc29a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.218733 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a232e198-17b4-4617-a664-b8db60b0d3fe","Type":"ContainerStarted","Data":"8921d43e85b1e84b6946bc3384c0ee2bac7aa783f0b5ed50513c9ee51fa1e0cd"} Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.220925 4677 generic.go:334] "Generic (PLEG): container finished" podID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerID="0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2" exitCode=0 Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.220982 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" event={"ID":"70265ca5-cffd-42a9-b519-5c5eed5fc29a","Type":"ContainerDied","Data":"0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2"} Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.221060 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" event={"ID":"70265ca5-cffd-42a9-b519-5c5eed5fc29a","Type":"ContainerDied","Data":"83891f189a8a979639ddc22605e74721ce2c8bad511362ef7a8630adc1b37cce"} Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.221062 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d946848f-qz6ws" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.221079 4677 scope.go:117] "RemoveContainer" containerID="0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2" Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.267971 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85d946848f-qz6ws"] Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.279836 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85d946848f-qz6ws"] Dec 03 14:08:54 crc kubenswrapper[4677]: I1203 14:08:54.313412 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70265ca5-cffd-42a9-b519-5c5eed5fc29a-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:54 crc kubenswrapper[4677]: W1203 14:08:54.488801 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cf253b5_7e86_4dfe_9b49_5fa2869b2473.slice/crio-af1aebb266fc6d336c596d4a841fc46490cb58aa3fdbbdb79780678eb8dde3fb WatchSource:0}: Error finding container af1aebb266fc6d336c596d4a841fc46490cb58aa3fdbbdb79780678eb8dde3fb: Status 404 returned error can't find the container with id af1aebb266fc6d336c596d4a841fc46490cb58aa3fdbbdb79780678eb8dde3fb Dec 03 14:08:54 crc kubenswrapper[4677]: W1203 14:08:54.501343 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67a0d4af_5eb2_42e1_b66d_a4f9d5e4c1c1.slice/crio-c9bd25776a752ac8fb8859f2f950c68077994ca38899f9d7966729470e6a0f54 WatchSource:0}: Error finding container c9bd25776a752ac8fb8859f2f950c68077994ca38899f9d7966729470e6a0f54: Status 404 returned error can't find the container with id c9bd25776a752ac8fb8859f2f950c68077994ca38899f9d7966729470e6a0f54 Dec 03 14:08:55 crc kubenswrapper[4677]: I1203 14:08:55.230394 4677 generic.go:334] "Generic (PLEG): container finished" podID="d39bf749-8b60-4833-97e0-b00791e11467" containerID="1986017e383b5d56c89a52892ee9f9ae17e8953c16a04f1d673fdff09b140110" exitCode=0 Dec 03 14:08:55 crc kubenswrapper[4677]: I1203 14:08:55.230511 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qqscx" event={"ID":"d39bf749-8b60-4833-97e0-b00791e11467","Type":"ContainerDied","Data":"1986017e383b5d56c89a52892ee9f9ae17e8953c16a04f1d673fdff09b140110"} Dec 03 14:08:55 crc kubenswrapper[4677]: I1203 14:08:55.231303 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-rh642" event={"ID":"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1","Type":"ContainerStarted","Data":"c9bd25776a752ac8fb8859f2f950c68077994ca38899f9d7966729470e6a0f54"} Dec 03 14:08:55 crc kubenswrapper[4677]: I1203 14:08:55.232805 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-6844-account-create-update-t6r4j" event={"ID":"7cf253b5-7e86-4dfe-9b49-5fa2869b2473","Type":"ContainerStarted","Data":"af1aebb266fc6d336c596d4a841fc46490cb58aa3fdbbdb79780678eb8dde3fb"} Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.035463 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" path="/var/lib/kubelet/pods/70265ca5-cffd-42a9-b519-5c5eed5fc29a/volumes" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.241742 4677 generic.go:334] "Generic (PLEG): container finished" podID="315da922-99be-44c7-81d8-24f0cf55490a" containerID="a0583341d41688452dc42a9a42e55645154003680f58e50ed9e609b979cbf91f" exitCode=0 Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.241799 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"315da922-99be-44c7-81d8-24f0cf55490a","Type":"ContainerDied","Data":"a0583341d41688452dc42a9a42e55645154003680f58e50ed9e609b979cbf91f"} Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.244363 4677 generic.go:334] "Generic (PLEG): container finished" podID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerID="c6d03f88aa96a8f0d3893995f7b8086caba5f3f4e8991faae35b447b28f9527a" exitCode=0 Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.244405 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7d5638b0-3c75-4290-b721-d2b5733a0cd3","Type":"ContainerDied","Data":"c6d03f88aa96a8f0d3893995f7b8086caba5f3f4e8991faae35b447b28f9527a"} Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.246970 4677 generic.go:334] "Generic (PLEG): container finished" podID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerID="52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038" exitCode=0 Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.247095 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ea48639b-8147-44bb-aa75-96c2ffc0921f","Type":"ContainerDied","Data":"52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038"} Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.270388 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jpkmm" podUID="f14465fa-42ae-4e85-bb6f-f2ef1bc5b681" containerName="ovn-controller" probeResult="failure" output=< Dec 03 14:08:56 crc kubenswrapper[4677]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 14:08:56 crc kubenswrapper[4677]: > Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.286658 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.291316 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-msfg2" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.556398 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jpkmm-config-7r6wp"] Dec 03 14:08:56 crc kubenswrapper[4677]: E1203 14:08:56.557169 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerName="dnsmasq-dns" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.557184 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerName="dnsmasq-dns" Dec 03 14:08:56 crc kubenswrapper[4677]: E1203 14:08:56.557218 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerName="init" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.557224 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerName="init" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.557569 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="70265ca5-cffd-42a9-b519-5c5eed5fc29a" containerName="dnsmasq-dns" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.558430 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.565973 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jpkmm-config-7r6wp"] Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.590194 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.694442 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.694614 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-scripts\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.694637 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrj5v\" (UniqueName: \"kubernetes.io/projected/e78c5f2f-d68e-42a5-9f7a-7768a7296319-kube-api-access-zrj5v\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.694673 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-additional-scripts\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.694690 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run-ovn\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.694712 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-log-ovn\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.795965 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.796073 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-scripts\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.796095 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrj5v\" (UniqueName: \"kubernetes.io/projected/e78c5f2f-d68e-42a5-9f7a-7768a7296319-kube-api-access-zrj5v\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.796131 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-additional-scripts\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.796145 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run-ovn\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.796166 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-log-ovn\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.796308 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.797134 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-log-ovn\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.798045 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run-ovn\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.800606 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-additional-scripts\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.804557 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-scripts\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.819065 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrj5v\" (UniqueName: \"kubernetes.io/projected/e78c5f2f-d68e-42a5-9f7a-7768a7296319-kube-api-access-zrj5v\") pod \"ovn-controller-jpkmm-config-7r6wp\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:56 crc kubenswrapper[4677]: I1203 14:08:56.888348 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.575563 4677 scope.go:117] "RemoveContainer" containerID="b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.768778 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.798090 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.798464 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.798630 4677 scope.go:117] "RemoveContainer" containerID="0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2" Dec 03 14:08:57 crc kubenswrapper[4677]: E1203 14:08:57.798897 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2\": container with ID starting with 0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2 not found: ID does not exist" containerID="0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.798937 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2"} err="failed to get container status \"0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2\": rpc error: code = NotFound desc = could not find container \"0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2\": container with ID starting with 0d9578d08009ea1605cfd6acaea103aa9f17ba851fb8c503709f6372002fe4b2 not found: ID does not exist" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.798974 4677 scope.go:117] "RemoveContainer" containerID="b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b" Dec 03 14:08:57 crc kubenswrapper[4677]: E1203 14:08:57.799220 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b\": container with ID starting with b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b not found: ID does not exist" containerID="b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.799243 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b"} err="failed to get container status \"b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b\": rpc error: code = NotFound desc = could not find container \"b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b\": container with ID starting with b25908c72b0395561d406694bfcffe05c78fc7131a2cb00422274f74f9430a4b not found: ID does not exist" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.806835 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.826324 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922226 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-operator-scripts\") pod \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922518 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79h9f\" (UniqueName: \"kubernetes.io/projected/445b8de4-6203-494a-9448-86f12765390e-kube-api-access-79h9f\") pod \"445b8de4-6203-494a-9448-86f12765390e\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922553 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2994ba2a-3d45-4303-b0c1-9800979e1cb5-operator-scripts\") pod \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922574 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-ring-data-devices\") pod \"d39bf749-8b60-4833-97e0-b00791e11467\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922606 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/445b8de4-6203-494a-9448-86f12765390e-operator-scripts\") pod \"445b8de4-6203-494a-9448-86f12765390e\" (UID: \"445b8de4-6203-494a-9448-86f12765390e\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922623 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z66sm\" (UniqueName: \"kubernetes.io/projected/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-kube-api-access-z66sm\") pod \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\" (UID: \"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922711 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldjjv\" (UniqueName: \"kubernetes.io/projected/2994ba2a-3d45-4303-b0c1-9800979e1cb5-kube-api-access-ldjjv\") pod \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\" (UID: \"2994ba2a-3d45-4303-b0c1-9800979e1cb5\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922733 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-swiftconf\") pod \"d39bf749-8b60-4833-97e0-b00791e11467\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922783 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-scripts\") pod \"d39bf749-8b60-4833-97e0-b00791e11467\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922799 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2bvv\" (UniqueName: \"kubernetes.io/projected/e7c96896-06fc-4875-a6c0-f0f3cec43b90-kube-api-access-k2bvv\") pod \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922833 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-combined-ca-bundle\") pod \"d39bf749-8b60-4833-97e0-b00791e11467\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922856 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d39bf749-8b60-4833-97e0-b00791e11467-etc-swift\") pod \"d39bf749-8b60-4833-97e0-b00791e11467\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922877 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qlfs\" (UniqueName: \"kubernetes.io/projected/d39bf749-8b60-4833-97e0-b00791e11467-kube-api-access-7qlfs\") pod \"d39bf749-8b60-4833-97e0-b00791e11467\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922913 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-dispersionconf\") pod \"d39bf749-8b60-4833-97e0-b00791e11467\" (UID: \"d39bf749-8b60-4833-97e0-b00791e11467\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.922961 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c96896-06fc-4875-a6c0-f0f3cec43b90-operator-scripts\") pod \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\" (UID: \"e7c96896-06fc-4875-a6c0-f0f3cec43b90\") " Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.923220 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/445b8de4-6203-494a-9448-86f12765390e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "445b8de4-6203-494a-9448-86f12765390e" (UID: "445b8de4-6203-494a-9448-86f12765390e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.923224 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2994ba2a-3d45-4303-b0c1-9800979e1cb5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2994ba2a-3d45-4303-b0c1-9800979e1cb5" (UID: "2994ba2a-3d45-4303-b0c1-9800979e1cb5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.923372 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2994ba2a-3d45-4303-b0c1-9800979e1cb5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.923384 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/445b8de4-6203-494a-9448-86f12765390e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.923929 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c96896-06fc-4875-a6c0-f0f3cec43b90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7c96896-06fc-4875-a6c0-f0f3cec43b90" (UID: "e7c96896-06fc-4875-a6c0-f0f3cec43b90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.924066 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d39bf749-8b60-4833-97e0-b00791e11467" (UID: "d39bf749-8b60-4833-97e0-b00791e11467"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.924467 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0" (UID: "5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.925112 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d39bf749-8b60-4833-97e0-b00791e11467-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d39bf749-8b60-4833-97e0-b00791e11467" (UID: "d39bf749-8b60-4833-97e0-b00791e11467"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.931374 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-kube-api-access-z66sm" (OuterVolumeSpecName: "kube-api-access-z66sm") pod "5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0" (UID: "5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0"). InnerVolumeSpecName "kube-api-access-z66sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.939208 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2994ba2a-3d45-4303-b0c1-9800979e1cb5-kube-api-access-ldjjv" (OuterVolumeSpecName: "kube-api-access-ldjjv") pod "2994ba2a-3d45-4303-b0c1-9800979e1cb5" (UID: "2994ba2a-3d45-4303-b0c1-9800979e1cb5"). InnerVolumeSpecName "kube-api-access-ldjjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.953277 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7c96896-06fc-4875-a6c0-f0f3cec43b90-kube-api-access-k2bvv" (OuterVolumeSpecName: "kube-api-access-k2bvv") pod "e7c96896-06fc-4875-a6c0-f0f3cec43b90" (UID: "e7c96896-06fc-4875-a6c0-f0f3cec43b90"). InnerVolumeSpecName "kube-api-access-k2bvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.953870 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445b8de4-6203-494a-9448-86f12765390e-kube-api-access-79h9f" (OuterVolumeSpecName: "kube-api-access-79h9f") pod "445b8de4-6203-494a-9448-86f12765390e" (UID: "445b8de4-6203-494a-9448-86f12765390e"). InnerVolumeSpecName "kube-api-access-79h9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.956417 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d39bf749-8b60-4833-97e0-b00791e11467-kube-api-access-7qlfs" (OuterVolumeSpecName: "kube-api-access-7qlfs") pod "d39bf749-8b60-4833-97e0-b00791e11467" (UID: "d39bf749-8b60-4833-97e0-b00791e11467"). InnerVolumeSpecName "kube-api-access-7qlfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.962363 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d39bf749-8b60-4833-97e0-b00791e11467" (UID: "d39bf749-8b60-4833-97e0-b00791e11467"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.963053 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d39bf749-8b60-4833-97e0-b00791e11467" (UID: "d39bf749-8b60-4833-97e0-b00791e11467"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.974355 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-scripts" (OuterVolumeSpecName: "scripts") pod "d39bf749-8b60-4833-97e0-b00791e11467" (UID: "d39bf749-8b60-4833-97e0-b00791e11467"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:08:57 crc kubenswrapper[4677]: I1203 14:08:57.994602 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d39bf749-8b60-4833-97e0-b00791e11467" (UID: "d39bf749-8b60-4833-97e0-b00791e11467"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025257 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldjjv\" (UniqueName: \"kubernetes.io/projected/2994ba2a-3d45-4303-b0c1-9800979e1cb5-kube-api-access-ldjjv\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025296 4677 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025307 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025318 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2bvv\" (UniqueName: \"kubernetes.io/projected/e7c96896-06fc-4875-a6c0-f0f3cec43b90-kube-api-access-k2bvv\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025330 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025340 4677 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d39bf749-8b60-4833-97e0-b00791e11467-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025350 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qlfs\" (UniqueName: \"kubernetes.io/projected/d39bf749-8b60-4833-97e0-b00791e11467-kube-api-access-7qlfs\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025361 4677 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d39bf749-8b60-4833-97e0-b00791e11467-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025371 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c96896-06fc-4875-a6c0-f0f3cec43b90-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025382 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025394 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79h9f\" (UniqueName: \"kubernetes.io/projected/445b8de4-6203-494a-9448-86f12765390e-kube-api-access-79h9f\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025406 4677 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d39bf749-8b60-4833-97e0-b00791e11467-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.025417 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z66sm\" (UniqueName: \"kubernetes.io/projected/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0-kube-api-access-z66sm\") on node \"crc\" DevicePath \"\"" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.270015 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ztpkm" event={"ID":"2994ba2a-3d45-4303-b0c1-9800979e1cb5","Type":"ContainerDied","Data":"1fd1a41dfafea211882e64c49588813a4d53544a0545971d7db425cef619efb6"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.270313 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fd1a41dfafea211882e64c49588813a4d53544a0545971d7db425cef619efb6" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.270380 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ztpkm" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.276038 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"315da922-99be-44c7-81d8-24f0cf55490a","Type":"ContainerStarted","Data":"cab514a3c3af4be479eb0bf2f8d0b0dd219397513696c94167a9031ba9ebc203"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.276345 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.279099 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-rh642" event={"ID":"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1","Type":"ContainerStarted","Data":"dbf149195267d8880ff1c49682426d35043703a40b9bf8166608aba2f0993b46"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.287209 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-da07-account-create-update-fpsng" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.287473 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-da07-account-create-update-fpsng" event={"ID":"445b8de4-6203-494a-9448-86f12765390e","Type":"ContainerDied","Data":"ed43f5f6f167d41458f1b25924c065581f3dad1a4d91912842ba83f2f2cbabae"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.287519 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed43f5f6f167d41458f1b25924c065581f3dad1a4d91912842ba83f2f2cbabae" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.290290 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7d5638b0-3c75-4290-b721-d2b5733a0cd3","Type":"ContainerStarted","Data":"4ecbb69a97190bfc1b6f0687cc33b90e8d85b0f157f94929cbcc8647bf0808fb"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.291423 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.294814 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jpkmm-config-7r6wp"] Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.300551 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-zk8qb" event={"ID":"e7c96896-06fc-4875-a6c0-f0f3cec43b90","Type":"ContainerDied","Data":"ea0f3b1f03f89c71d5b9c517dbd21a4744ad6d91b990402266437c755ae39e7b"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.300598 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea0f3b1f03f89c71d5b9c517dbd21a4744ad6d91b990402266437c755ae39e7b" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.300567 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zk8qb" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.315593 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qqscx" event={"ID":"d39bf749-8b60-4833-97e0-b00791e11467","Type":"ContainerDied","Data":"193a2684c99c670c8bd63dcc776127459cafe6868e0e7e95ca4359919b806221"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.315638 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="193a2684c99c670c8bd63dcc776127459cafe6868e0e7e95ca4359919b806221" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.315664 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qqscx" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.322324 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=56.584166801 podStartE2EDuration="1m13.32230255s" podCreationTimestamp="2025-12-03 14:07:45 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.082211446 +0000 UTC m=+1275.828543901" lastFinishedPulling="2025-12-03 14:08:21.820347185 +0000 UTC m=+1292.566679650" observedRunningTime="2025-12-03 14:08:58.314008998 +0000 UTC m=+1329.060341473" watchObservedRunningTime="2025-12-03 14:08:58.32230255 +0000 UTC m=+1329.068635005" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.334981 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bbef-account-create-update-8btps" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.334977 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bbef-account-create-update-8btps" event={"ID":"5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0","Type":"ContainerDied","Data":"f6b4aa7bdff2dd6fc3c6f491c6753be7afd2c4f13c9c84af6125ed22984808ef"} Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.335856 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6b4aa7bdff2dd6fc3c6f491c6753be7afd2c4f13c9c84af6125ed22984808ef" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.350915 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-create-rh642" podStartSLOduration=6.350897211 podStartE2EDuration="6.350897211s" podCreationTimestamp="2025-12-03 14:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:58.343814091 +0000 UTC m=+1329.090146546" watchObservedRunningTime="2025-12-03 14:08:58.350897211 +0000 UTC m=+1329.097229666" Dec 03 14:08:58 crc kubenswrapper[4677]: I1203 14:08:58.405034 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371963.449766 podStartE2EDuration="1m13.405009965s" podCreationTimestamp="2025-12-03 14:07:45 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.071805716 +0000 UTC m=+1275.818138181" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:58.391857559 +0000 UTC m=+1329.138190024" watchObservedRunningTime="2025-12-03 14:08:58.405009965 +0000 UTC m=+1329.151342410" Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.372787 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ea48639b-8147-44bb-aa75-96c2ffc0921f","Type":"ContainerStarted","Data":"2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.373252 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.374848 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerStarted","Data":"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.376445 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jpkmm-config-7r6wp" event={"ID":"e78c5f2f-d68e-42a5-9f7a-7768a7296319","Type":"ContainerStarted","Data":"c3e5a3fab9ee33457e644fbf29a56ed57ceb2fc629f3a944fa9b2cf28c4d4933"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.376529 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jpkmm-config-7r6wp" event={"ID":"e78c5f2f-d68e-42a5-9f7a-7768a7296319","Type":"ContainerStarted","Data":"df13fb959d0aa673f4bcd4bfe617e8fa57cde97db897e51fd532ec02cca7f650"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.377841 4677 generic.go:334] "Generic (PLEG): container finished" podID="67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1" containerID="dbf149195267d8880ff1c49682426d35043703a40b9bf8166608aba2f0993b46" exitCode=0 Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.377900 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-rh642" event={"ID":"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1","Type":"ContainerDied","Data":"dbf149195267d8880ff1c49682426d35043703a40b9bf8166608aba2f0993b46"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.379749 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a232e198-17b4-4617-a664-b8db60b0d3fe","Type":"ContainerStarted","Data":"c5f4b09eb35f73a7a9bc769da564c0042dd3cdcd20ef7ec526c710022a336c01"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.379786 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a232e198-17b4-4617-a664-b8db60b0d3fe","Type":"ContainerStarted","Data":"33de9c913458cd55be4e3e8682684726e1329b7264f06cb4f2ba247f07b3b5b3"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.381375 4677 generic.go:334] "Generic (PLEG): container finished" podID="7cf253b5-7e86-4dfe-9b49-5fa2869b2473" containerID="44a8375ab0ed916c8d8c90009b485a19afb22f436de367ac58b3af7b022fb996" exitCode=0 Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.381421 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-6844-account-create-update-t6r4j" event={"ID":"7cf253b5-7e86-4dfe-9b49-5fa2869b2473","Type":"ContainerDied","Data":"44a8375ab0ed916c8d8c90009b485a19afb22f436de367ac58b3af7b022fb996"} Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.402968 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=57.303529792 podStartE2EDuration="1m13.40293998s" podCreationTimestamp="2025-12-03 14:07:46 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.480348276 +0000 UTC m=+1276.226680731" lastFinishedPulling="2025-12-03 14:08:21.579758464 +0000 UTC m=+1292.326090919" observedRunningTime="2025-12-03 14:08:59.397594474 +0000 UTC m=+1330.143926929" watchObservedRunningTime="2025-12-03 14:08:59.40293998 +0000 UTC m=+1330.149272435" Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.422072 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.921802973 podStartE2EDuration="7.422057259s" podCreationTimestamp="2025-12-03 14:08:52 +0000 UTC" firstStartedPulling="2025-12-03 14:08:53.202251702 +0000 UTC m=+1323.948584157" lastFinishedPulling="2025-12-03 14:08:57.702505978 +0000 UTC m=+1328.448838443" observedRunningTime="2025-12-03 14:08:59.415402759 +0000 UTC m=+1330.161735214" watchObservedRunningTime="2025-12-03 14:08:59.422057259 +0000 UTC m=+1330.168389714" Dec 03 14:08:59 crc kubenswrapper[4677]: I1203 14:08:59.462214 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jpkmm-config-7r6wp" podStartSLOduration=3.462198926 podStartE2EDuration="3.462198926s" podCreationTimestamp="2025-12-03 14:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:08:59.458169122 +0000 UTC m=+1330.204501567" watchObservedRunningTime="2025-12-03 14:08:59.462198926 +0000 UTC m=+1330.208531381" Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.394412 4677 generic.go:334] "Generic (PLEG): container finished" podID="e78c5f2f-d68e-42a5-9f7a-7768a7296319" containerID="c3e5a3fab9ee33457e644fbf29a56ed57ceb2fc629f3a944fa9b2cf28c4d4933" exitCode=0 Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.395583 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jpkmm-config-7r6wp" event={"ID":"e78c5f2f-d68e-42a5-9f7a-7768a7296319","Type":"ContainerDied","Data":"c3e5a3fab9ee33457e644fbf29a56ed57ceb2fc629f3a944fa9b2cf28c4d4933"} Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.396875 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.880030 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-rh642" Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.884575 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.973035 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtfqh\" (UniqueName: \"kubernetes.io/projected/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-kube-api-access-jtfqh\") pod \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.973183 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-operator-scripts\") pod \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\" (UID: \"7cf253b5-7e86-4dfe-9b49-5fa2869b2473\") " Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.973330 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-operator-scripts\") pod \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.973416 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxrbm\" (UniqueName: \"kubernetes.io/projected/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-kube-api-access-zxrbm\") pod \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\" (UID: \"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1\") " Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.975381 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1" (UID: "67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.975411 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7cf253b5-7e86-4dfe-9b49-5fa2869b2473" (UID: "7cf253b5-7e86-4dfe-9b49-5fa2869b2473"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.981957 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-kube-api-access-zxrbm" (OuterVolumeSpecName: "kube-api-access-zxrbm") pod "67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1" (UID: "67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1"). InnerVolumeSpecName "kube-api-access-zxrbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:00 crc kubenswrapper[4677]: I1203 14:09:00.988366 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-kube-api-access-jtfqh" (OuterVolumeSpecName: "kube-api-access-jtfqh") pod "7cf253b5-7e86-4dfe-9b49-5fa2869b2473" (UID: "7cf253b5-7e86-4dfe-9b49-5fa2869b2473"). InnerVolumeSpecName "kube-api-access-jtfqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.077265 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtfqh\" (UniqueName: \"kubernetes.io/projected/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-kube-api-access-jtfqh\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.077311 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf253b5-7e86-4dfe-9b49-5fa2869b2473-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.077322 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.077332 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxrbm\" (UniqueName: \"kubernetes.io/projected/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1-kube-api-access-zxrbm\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.277769 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jpkmm" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.402632 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-rh642" event={"ID":"67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1","Type":"ContainerDied","Data":"c9bd25776a752ac8fb8859f2f950c68077994ca38899f9d7966729470e6a0f54"} Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.402669 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9bd25776a752ac8fb8859f2f950c68077994ca38899f9d7966729470e6a0f54" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.402728 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-rh642" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.405006 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-6844-account-create-update-t6r4j" event={"ID":"7cf253b5-7e86-4dfe-9b49-5fa2869b2473","Type":"ContainerDied","Data":"af1aebb266fc6d336c596d4a841fc46490cb58aa3fdbbdb79780678eb8dde3fb"} Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.405043 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af1aebb266fc6d336c596d4a841fc46490cb58aa3fdbbdb79780678eb8dde3fb" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.405098 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-6844-account-create-update-t6r4j" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.697002 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.787301 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-scripts\") pod \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.787355 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run-ovn\") pod \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.787396 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-log-ovn\") pod \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.787425 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run\") pod \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.787463 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-additional-scripts\") pod \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.787511 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrj5v\" (UniqueName: \"kubernetes.io/projected/e78c5f2f-d68e-42a5-9f7a-7768a7296319-kube-api-access-zrj5v\") pod \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\" (UID: \"e78c5f2f-d68e-42a5-9f7a-7768a7296319\") " Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.788740 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e78c5f2f-d68e-42a5-9f7a-7768a7296319" (UID: "e78c5f2f-d68e-42a5-9f7a-7768a7296319"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.789405 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run" (OuterVolumeSpecName: "var-run") pod "e78c5f2f-d68e-42a5-9f7a-7768a7296319" (UID: "e78c5f2f-d68e-42a5-9f7a-7768a7296319"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.789463 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e78c5f2f-d68e-42a5-9f7a-7768a7296319" (UID: "e78c5f2f-d68e-42a5-9f7a-7768a7296319"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.790076 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e78c5f2f-d68e-42a5-9f7a-7768a7296319" (UID: "e78c5f2f-d68e-42a5-9f7a-7768a7296319"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.794174 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78c5f2f-d68e-42a5-9f7a-7768a7296319-kube-api-access-zrj5v" (OuterVolumeSpecName: "kube-api-access-zrj5v") pod "e78c5f2f-d68e-42a5-9f7a-7768a7296319" (UID: "e78c5f2f-d68e-42a5-9f7a-7768a7296319"). InnerVolumeSpecName "kube-api-access-zrj5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.894748 4677 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.894799 4677 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.894811 4677 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e78c5f2f-d68e-42a5-9f7a-7768a7296319-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.894821 4677 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.894835 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrj5v\" (UniqueName: \"kubernetes.io/projected/e78c5f2f-d68e-42a5-9f7a-7768a7296319-kube-api-access-zrj5v\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.928426 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-scripts" (OuterVolumeSpecName: "scripts") pod "e78c5f2f-d68e-42a5-9f7a-7768a7296319" (UID: "e78c5f2f-d68e-42a5-9f7a-7768a7296319"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:01 crc kubenswrapper[4677]: I1203 14:09:01.996297 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e78c5f2f-d68e-42a5-9f7a-7768a7296319-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:02 crc kubenswrapper[4677]: I1203 14:09:02.414851 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jpkmm-config-7r6wp" event={"ID":"e78c5f2f-d68e-42a5-9f7a-7768a7296319","Type":"ContainerDied","Data":"df13fb959d0aa673f4bcd4bfe617e8fa57cde97db897e51fd532ec02cca7f650"} Dec 03 14:09:02 crc kubenswrapper[4677]: I1203 14:09:02.415188 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df13fb959d0aa673f4bcd4bfe617e8fa57cde97db897e51fd532ec02cca7f650" Dec 03 14:09:02 crc kubenswrapper[4677]: I1203 14:09:02.414897 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jpkmm-config-7r6wp" Dec 03 14:09:02 crc kubenswrapper[4677]: I1203 14:09:02.419549 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerStarted","Data":"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50"} Dec 03 14:09:02 crc kubenswrapper[4677]: I1203 14:09:02.824086 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jpkmm-config-7r6wp"] Dec 03 14:09:02 crc kubenswrapper[4677]: I1203 14:09:02.832354 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jpkmm-config-7r6wp"] Dec 03 14:09:03 crc kubenswrapper[4677]: I1203 14:09:03.985250 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78c5f2f-d68e-42a5-9f7a-7768a7296319" path="/var/lib/kubelet/pods/e78c5f2f-d68e-42a5-9f7a-7768a7296319/volumes" Dec 03 14:09:06 crc kubenswrapper[4677]: I1203 14:09:06.384778 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:09:06 crc kubenswrapper[4677]: I1203 14:09:06.405272 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b25cf219-67f3-4764-a5da-3dbe89f71641-etc-swift\") pod \"swift-storage-0\" (UID: \"b25cf219-67f3-4764-a5da-3dbe89f71641\") " pod="openstack/swift-storage-0" Dec 03 14:09:06 crc kubenswrapper[4677]: I1203 14:09:06.509872 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 14:09:07 crc kubenswrapper[4677]: I1203 14:09:07.463147 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerStarted","Data":"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3"} Dec 03 14:09:07 crc kubenswrapper[4677]: I1203 14:09:07.495749 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=13.966277082 podStartE2EDuration="1m15.495728872s" podCreationTimestamp="2025-12-03 14:07:52 +0000 UTC" firstStartedPulling="2025-12-03 14:08:05.532046936 +0000 UTC m=+1276.278379391" lastFinishedPulling="2025-12-03 14:09:07.061498716 +0000 UTC m=+1337.807831181" observedRunningTime="2025-12-03 14:09:07.484361102 +0000 UTC m=+1338.230693577" watchObservedRunningTime="2025-12-03 14:09:07.495728872 +0000 UTC m=+1338.242061327" Dec 03 14:09:07 crc kubenswrapper[4677]: W1203 14:09:07.628421 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb25cf219_67f3_4764_a5da_3dbe89f71641.slice/crio-849c63323260a3b029b748879c089cf3ba524c6d959dad79451dab28277eabf8 WatchSource:0}: Error finding container 849c63323260a3b029b748879c089cf3ba524c6d959dad79451dab28277eabf8: Status 404 returned error can't find the container with id 849c63323260a3b029b748879c089cf3ba524c6d959dad79451dab28277eabf8 Dec 03 14:09:07 crc kubenswrapper[4677]: I1203 14:09:07.631863 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 14:09:08 crc kubenswrapper[4677]: I1203 14:09:08.442933 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:09:08 crc kubenswrapper[4677]: I1203 14:09:08.443873 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:09:08 crc kubenswrapper[4677]: I1203 14:09:08.472920 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"bcbab55bb67de508a856f58236644d4a46167a07802c3d11cf03c1ba2f066320"} Dec 03 14:09:08 crc kubenswrapper[4677]: I1203 14:09:08.473006 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"849c63323260a3b029b748879c089cf3ba524c6d959dad79451dab28277eabf8"} Dec 03 14:09:09 crc kubenswrapper[4677]: I1203 14:09:09.331164 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:09 crc kubenswrapper[4677]: I1203 14:09:09.331849 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:09 crc kubenswrapper[4677]: I1203 14:09:09.333931 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:09 crc kubenswrapper[4677]: I1203 14:09:09.484935 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"6742b760243bce85761c92c2f95224185854dcb9da5546ed8120174ba42048d1"} Dec 03 14:09:09 crc kubenswrapper[4677]: I1203 14:09:09.485291 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"ea35a97be73810de019285d3f1fbc3750481d779d919b88449f943e6a8b5581d"} Dec 03 14:09:09 crc kubenswrapper[4677]: I1203 14:09:09.485304 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"f048a9eca03dd4f926aba7f789e32888a25c114701b3a8ef6c20ae23ff8cf85c"} Dec 03 14:09:09 crc kubenswrapper[4677]: I1203 14:09:09.486038 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:10 crc kubenswrapper[4677]: I1203 14:09:10.504444 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"3536111e3c5210a8c8b5a6d7aedc6abf70a4c690d92e9e12ef6d6d3ef7323eb2"} Dec 03 14:09:11 crc kubenswrapper[4677]: I1203 14:09:11.536212 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"5bb4aac83ad064d6fb745c7503c95b176ffd7fa2b3e6dbe73d9fe7af48aa8035"} Dec 03 14:09:11 crc kubenswrapper[4677]: I1203 14:09:11.536573 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"5f9afcf71d4960133e0cab50215e382afe474365d31ef91e75276ea2127817b9"} Dec 03 14:09:11 crc kubenswrapper[4677]: I1203 14:09:11.536587 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"10b49739da24122d9a494f230bd567694d449af9395cbdbb238e8cfe5045e46c"} Dec 03 14:09:11 crc kubenswrapper[4677]: I1203 14:09:11.851552 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:09:12 crc kubenswrapper[4677]: I1203 14:09:12.551331 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"5825b8cbaa6f09de006ea1da3afa89cb0110ebce1711d7bc5db5b28b9b84e6fa"} Dec 03 14:09:12 crc kubenswrapper[4677]: I1203 14:09:12.551786 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="config-reloader" containerID="cri-o://d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50" gracePeriod=600 Dec 03 14:09:12 crc kubenswrapper[4677]: I1203 14:09:12.551717 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="thanos-sidecar" containerID="cri-o://d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3" gracePeriod=600 Dec 03 14:09:12 crc kubenswrapper[4677]: I1203 14:09:12.551809 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"85ec77f443d5d973a8dfe8b94ea14cfab00616131bab0a9e3c5ef57557e0d40f"} Dec 03 14:09:12 crc kubenswrapper[4677]: I1203 14:09:12.551977 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"b6fa0bac0af906d4774179bdd96390bc5b66c23e3f3dd63265d10a3753e373ee"} Dec 03 14:09:12 crc kubenswrapper[4677]: I1203 14:09:12.551586 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="prometheus" containerID="cri-o://fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd" gracePeriod=600 Dec 03 14:09:12 crc kubenswrapper[4677]: I1203 14:09:12.814016 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.528425 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.573460 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"24f985ab2b5f4edfbede00789bac16dcbcae0c32551178a40383a7cb4417f135"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.573502 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"f4e335dbe0c977dc34a9b07d7e2f514ef1871a46968bc05b66588d17fc1d0636"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.573511 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"0b6ec50ea7d1542ca435958b4987190c243307e7cde01a2fa0667f6b5c6438e3"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.573519 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b25cf219-67f3-4764-a5da-3dbe89f71641","Type":"ContainerStarted","Data":"f3b71e440c0c2e00ca7d99aa7cdc96e190db5f21808ea15bf503cfe33c84667d"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579278 4677 generic.go:334] "Generic (PLEG): container finished" podID="db9e192a-61af-41fa-882c-81879f2a556a" containerID="d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3" exitCode=0 Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579308 4677 generic.go:334] "Generic (PLEG): container finished" podID="db9e192a-61af-41fa-882c-81879f2a556a" containerID="d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50" exitCode=0 Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579317 4677 generic.go:334] "Generic (PLEG): container finished" podID="db9e192a-61af-41fa-882c-81879f2a556a" containerID="fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd" exitCode=0 Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579338 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerDied","Data":"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579363 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerDied","Data":"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579373 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerDied","Data":"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579395 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"db9e192a-61af-41fa-882c-81879f2a556a","Type":"ContainerDied","Data":"2ac15fca35376f76ef2d15786dfa90bf4a4904216e15d5ecf5f79fce40752e7d"} Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579410 4677 scope.go:117] "RemoveContainer" containerID="d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.579527 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603008 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-web-config\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603217 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603241 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/db9e192a-61af-41fa-882c-81879f2a556a-prometheus-metric-storage-rulefiles-0\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603262 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-thanos-prometheus-http-client-file\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603377 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-tls-assets\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603435 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-config\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603466 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w222\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-kube-api-access-6w222\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.603519 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/db9e192a-61af-41fa-882c-81879f2a556a-config-out\") pod \"db9e192a-61af-41fa-882c-81879f2a556a\" (UID: \"db9e192a-61af-41fa-882c-81879f2a556a\") " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.609314 4677 scope.go:117] "RemoveContainer" containerID="d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.612268 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db9e192a-61af-41fa-882c-81879f2a556a-config-out" (OuterVolumeSpecName: "config-out") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.619564 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db9e192a-61af-41fa-882c-81879f2a556a-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.623878 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.624334 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-config" (OuterVolumeSpecName: "config") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.625292 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-kube-api-access-6w222" (OuterVolumeSpecName: "kube-api-access-6w222") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "kube-api-access-6w222". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.630210 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.653784 4677 scope.go:117] "RemoveContainer" containerID="fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.653889 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.681808 4677 scope.go:117] "RemoveContainer" containerID="dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.681818 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-web-config" (OuterVolumeSpecName: "web-config") pod "db9e192a-61af-41fa-882c-81879f2a556a" (UID: "db9e192a-61af-41fa-882c-81879f2a556a"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706101 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706156 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w222\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-kube-api-access-6w222\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706170 4677 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/db9e192a-61af-41fa-882c-81879f2a556a-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706183 4677 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706242 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") on node \"crc\" " Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706259 4677 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/db9e192a-61af-41fa-882c-81879f2a556a-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706298 4677 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/db9e192a-61af-41fa-882c-81879f2a556a-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.706315 4677 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/db9e192a-61af-41fa-882c-81879f2a556a-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.708159 4677 scope.go:117] "RemoveContainer" containerID="d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.710299 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3\": container with ID starting with d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3 not found: ID does not exist" containerID="d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.710331 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3"} err="failed to get container status \"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3\": rpc error: code = NotFound desc = could not find container \"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3\": container with ID starting with d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3 not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.710353 4677 scope.go:117] "RemoveContainer" containerID="d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.711543 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50\": container with ID starting with d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50 not found: ID does not exist" containerID="d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.711617 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50"} err="failed to get container status \"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50\": rpc error: code = NotFound desc = could not find container \"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50\": container with ID starting with d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50 not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.711651 4677 scope.go:117] "RemoveContainer" containerID="fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.712485 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd\": container with ID starting with fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd not found: ID does not exist" containerID="fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.712516 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd"} err="failed to get container status \"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd\": rpc error: code = NotFound desc = could not find container \"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd\": container with ID starting with fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.712537 4677 scope.go:117] "RemoveContainer" containerID="dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.713084 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb\": container with ID starting with dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb not found: ID does not exist" containerID="dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.713167 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb"} err="failed to get container status \"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb\": rpc error: code = NotFound desc = could not find container \"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb\": container with ID starting with dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.713241 4677 scope.go:117] "RemoveContainer" containerID="d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.713936 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3"} err="failed to get container status \"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3\": rpc error: code = NotFound desc = could not find container \"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3\": container with ID starting with d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3 not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.713979 4677 scope.go:117] "RemoveContainer" containerID="d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.715423 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50"} err="failed to get container status \"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50\": rpc error: code = NotFound desc = could not find container \"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50\": container with ID starting with d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50 not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.715441 4677 scope.go:117] "RemoveContainer" containerID="fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.715796 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd"} err="failed to get container status \"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd\": rpc error: code = NotFound desc = could not find container \"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd\": container with ID starting with fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.715813 4677 scope.go:117] "RemoveContainer" containerID="dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716128 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb"} err="failed to get container status \"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb\": rpc error: code = NotFound desc = could not find container \"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb\": container with ID starting with dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716145 4677 scope.go:117] "RemoveContainer" containerID="d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716364 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3"} err="failed to get container status \"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3\": rpc error: code = NotFound desc = could not find container \"d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3\": container with ID starting with d509bc9006e2bd28f15ebc46dae6bd1e7f23542eebd55039a3fcdd7a778509a3 not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716380 4677 scope.go:117] "RemoveContainer" containerID="d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716590 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50"} err="failed to get container status \"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50\": rpc error: code = NotFound desc = could not find container \"d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50\": container with ID starting with d2622d1be4865b845257932f0821b35ebf3f5cde545857317bf76d60596f2a50 not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716610 4677 scope.go:117] "RemoveContainer" containerID="fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716824 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd"} err="failed to get container status \"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd\": rpc error: code = NotFound desc = could not find container \"fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd\": container with ID starting with fe3ede8472943b3844e94f62b4fd341974598127bb3e64e95e51b3266c3574cd not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.716843 4677 scope.go:117] "RemoveContainer" containerID="dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.717180 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb"} err="failed to get container status \"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb\": rpc error: code = NotFound desc = could not find container \"dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb\": container with ID starting with dbe81a485e942aa94ed03a38936a90d313ea648cf77fd3a71d7597143e7bc5fb not found: ID does not exist" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.736786 4677 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.736973 4677 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208") on node "crc" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.807380 4677 reconciler_common.go:293] "Volume detached for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.932873 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.889733986 podStartE2EDuration="40.932853968s" podCreationTimestamp="2025-12-03 14:08:33 +0000 UTC" firstStartedPulling="2025-12-03 14:09:07.631074645 +0000 UTC m=+1338.377407100" lastFinishedPulling="2025-12-03 14:09:11.674194627 +0000 UTC m=+1342.420527082" observedRunningTime="2025-12-03 14:09:13.618063777 +0000 UTC m=+1344.364396242" watchObservedRunningTime="2025-12-03 14:09:13.932853968 +0000 UTC m=+1344.679186423" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.933702 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.947494 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965058 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj"] Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965440 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965463 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965478 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78c5f2f-d68e-42a5-9f7a-7768a7296319" containerName="ovn-config" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965484 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78c5f2f-d68e-42a5-9f7a-7768a7296319" containerName="ovn-config" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965498 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf253b5-7e86-4dfe-9b49-5fa2869b2473" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965503 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf253b5-7e86-4dfe-9b49-5fa2869b2473" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965513 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="thanos-sidecar" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965518 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="thanos-sidecar" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965529 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39bf749-8b60-4833-97e0-b00791e11467" containerName="swift-ring-rebalance" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965535 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39bf749-8b60-4833-97e0-b00791e11467" containerName="swift-ring-rebalance" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965546 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="config-reloader" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965551 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="config-reloader" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965561 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2994ba2a-3d45-4303-b0c1-9800979e1cb5" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965566 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="2994ba2a-3d45-4303-b0c1-9800979e1cb5" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965579 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="init-config-reloader" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965584 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="init-config-reloader" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965594 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c96896-06fc-4875-a6c0-f0f3cec43b90" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965599 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c96896-06fc-4875-a6c0-f0f3cec43b90" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965608 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445b8de4-6203-494a-9448-86f12765390e" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965614 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="445b8de4-6203-494a-9448-86f12765390e" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965624 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965631 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: E1203 14:09:13.965638 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="prometheus" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965645 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="prometheus" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965805 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="2994ba2a-3d45-4303-b0c1-9800979e1cb5" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965819 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965827 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="prometheus" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965842 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="thanos-sidecar" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965851 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf253b5-7e86-4dfe-9b49-5fa2869b2473" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965862 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965872 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="445b8de4-6203-494a-9448-86f12765390e" containerName="mariadb-account-create-update" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965879 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78c5f2f-d68e-42a5-9f7a-7768a7296319" containerName="ovn-config" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965888 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="d39bf749-8b60-4833-97e0-b00791e11467" containerName="swift-ring-rebalance" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965898 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7c96896-06fc-4875-a6c0-f0f3cec43b90" containerName="mariadb-database-create" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.965907 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="db9e192a-61af-41fa-882c-81879f2a556a" containerName="config-reloader" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.966920 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.968503 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.987131 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db9e192a-61af-41fa-882c-81879f2a556a" path="/var/lib/kubelet/pods/db9e192a-61af-41fa-882c-81879f2a556a/volumes" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.987987 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.990782 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.993973 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.994138 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mq9g8" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.995078 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.995285 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.998981 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 14:09:13 crc kubenswrapper[4677]: I1203 14:09:13.999339 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj"] Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.000252 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.003305 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.096629 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.114872 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4mtf\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-kube-api-access-k4mtf\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.114929 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt5xq\" (UniqueName: \"kubernetes.io/projected/3860befc-70ca-4658-b967-b3e8351c9788-kube-api-access-kt5xq\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.114988 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115016 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-swift-storage-0\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115090 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115185 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-sb\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115302 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-svc\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115372 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115432 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115503 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115612 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115665 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-nb\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115697 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115746 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115787 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115817 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-config\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.115863 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.216999 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-config\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217075 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217129 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4mtf\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-kube-api-access-k4mtf\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217147 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt5xq\" (UniqueName: \"kubernetes.io/projected/3860befc-70ca-4658-b967-b3e8351c9788-kube-api-access-kt5xq\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217173 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217189 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-swift-storage-0\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217205 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217225 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-sb\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217260 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-svc\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217283 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217317 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217344 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217369 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217389 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-nb\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217408 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217431 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.217452 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.218004 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-config\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.218733 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-svc\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.219531 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-swift-storage-0\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.219896 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-nb\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.220786 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.221398 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-sb\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.221423 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.222417 4677 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.222474 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.222474 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b3062aca7e122ea7b9aa119622cdce8d97263fccd877be8c6301bd417d0b47b0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.222778 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.223035 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.223201 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.223217 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.229169 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.229264 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.244743 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4mtf\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-kube-api-access-k4mtf\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.248080 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt5xq\" (UniqueName: \"kubernetes.io/projected/3860befc-70ca-4658-b967-b3e8351c9788-kube-api-access-kt5xq\") pod \"dnsmasq-dns-6cbbb6d9d7-5z4fj\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.277514 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.289184 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.342070 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.768427 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj"] Dec 03 14:09:14 crc kubenswrapper[4677]: I1203 14:09:14.887360 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:09:14 crc kubenswrapper[4677]: W1203 14:09:14.894197 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedfdba14_4c6d_4708_b024_2f5c7bc2f01b.slice/crio-f8276e27efd4f4de3a99f9723cf8afde1aaf1cf37781d7cb7413fd4b878f591c WatchSource:0}: Error finding container f8276e27efd4f4de3a99f9723cf8afde1aaf1cf37781d7cb7413fd4b878f591c: Status 404 returned error can't find the container with id f8276e27efd4f4de3a99f9723cf8afde1aaf1cf37781d7cb7413fd4b878f591c Dec 03 14:09:15 crc kubenswrapper[4677]: I1203 14:09:15.611716 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerStarted","Data":"f8276e27efd4f4de3a99f9723cf8afde1aaf1cf37781d7cb7413fd4b878f591c"} Dec 03 14:09:15 crc kubenswrapper[4677]: I1203 14:09:15.613663 4677 generic.go:334] "Generic (PLEG): container finished" podID="3860befc-70ca-4658-b967-b3e8351c9788" containerID="4f1344de56e500782cf8e8b798c387a14aa88f6e84cb7cd399436ae56ddce1d2" exitCode=0 Dec 03 14:09:15 crc kubenswrapper[4677]: I1203 14:09:15.613857 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" event={"ID":"3860befc-70ca-4658-b967-b3e8351c9788","Type":"ContainerDied","Data":"4f1344de56e500782cf8e8b798c387a14aa88f6e84cb7cd399436ae56ddce1d2"} Dec 03 14:09:15 crc kubenswrapper[4677]: I1203 14:09:15.614037 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" event={"ID":"3860befc-70ca-4658-b967-b3e8351c9788","Type":"ContainerStarted","Data":"a0f2903ad0a5471038439b36c4b502fc1ed8e5184b388bf8079e0e6be436ffe4"} Dec 03 14:09:16 crc kubenswrapper[4677]: I1203 14:09:16.626297 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" event={"ID":"3860befc-70ca-4658-b967-b3e8351c9788","Type":"ContainerStarted","Data":"ae5ab29a0972e40bec375a1634edb74351d79e1fd5e1d4ec70437460e2232b5c"} Dec 03 14:09:16 crc kubenswrapper[4677]: I1203 14:09:16.626981 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:16 crc kubenswrapper[4677]: I1203 14:09:16.657911 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" podStartSLOduration=3.657893578 podStartE2EDuration="3.657893578s" podCreationTimestamp="2025-12-03 14:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:16.656583085 +0000 UTC m=+1347.402915540" watchObservedRunningTime="2025-12-03 14:09:16.657893578 +0000 UTC m=+1347.404226053" Dec 03 14:09:17 crc kubenswrapper[4677]: I1203 14:09:17.325669 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Dec 03 14:09:17 crc kubenswrapper[4677]: I1203 14:09:17.364025 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="315da922-99be-44c7-81d8-24f0cf55490a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Dec 03 14:09:17 crc kubenswrapper[4677]: I1203 14:09:17.638389 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerStarted","Data":"131b5669e7bdb8fec5a14c35b257c557a9085a95a7fe9a66bf08a18a6b349756"} Dec 03 14:09:17 crc kubenswrapper[4677]: I1203 14:09:17.707481 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.106:5671: connect: connection refused" Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.291222 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.361800 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c98758bb9-jwrht"] Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.362287 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" podUID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerName="dnsmasq-dns" containerID="cri-o://1d923eb357efa003508d4e717adee4aa5a17e8572cf26b077ee79b1c77ca8dd8" gracePeriod=10 Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.706253 4677 generic.go:334] "Generic (PLEG): container finished" podID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerID="1d923eb357efa003508d4e717adee4aa5a17e8572cf26b077ee79b1c77ca8dd8" exitCode=0 Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.706332 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" event={"ID":"e0ebab4b-62db-4b77-b730-0413b51bdeaf","Type":"ContainerDied","Data":"1d923eb357efa003508d4e717adee4aa5a17e8572cf26b077ee79b1c77ca8dd8"} Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.798066 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.899986 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-sb\") pod \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.900471 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-config\") pod \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.900527 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkt8t\" (UniqueName: \"kubernetes.io/projected/e0ebab4b-62db-4b77-b730-0413b51bdeaf-kube-api-access-fkt8t\") pod \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.900600 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-nb\") pod \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.900709 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-dns-svc\") pod \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\" (UID: \"e0ebab4b-62db-4b77-b730-0413b51bdeaf\") " Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.905812 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ebab4b-62db-4b77-b730-0413b51bdeaf-kube-api-access-fkt8t" (OuterVolumeSpecName: "kube-api-access-fkt8t") pod "e0ebab4b-62db-4b77-b730-0413b51bdeaf" (UID: "e0ebab4b-62db-4b77-b730-0413b51bdeaf"). InnerVolumeSpecName "kube-api-access-fkt8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.949306 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e0ebab4b-62db-4b77-b730-0413b51bdeaf" (UID: "e0ebab4b-62db-4b77-b730-0413b51bdeaf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.953442 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e0ebab4b-62db-4b77-b730-0413b51bdeaf" (UID: "e0ebab4b-62db-4b77-b730-0413b51bdeaf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.958077 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-config" (OuterVolumeSpecName: "config") pod "e0ebab4b-62db-4b77-b730-0413b51bdeaf" (UID: "e0ebab4b-62db-4b77-b730-0413b51bdeaf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:24 crc kubenswrapper[4677]: I1203 14:09:24.962971 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e0ebab4b-62db-4b77-b730-0413b51bdeaf" (UID: "e0ebab4b-62db-4b77-b730-0413b51bdeaf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.002772 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.002809 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkt8t\" (UniqueName: \"kubernetes.io/projected/e0ebab4b-62db-4b77-b730-0413b51bdeaf-kube-api-access-fkt8t\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.002824 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.002838 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.002849 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0ebab4b-62db-4b77-b730-0413b51bdeaf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.717942 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" event={"ID":"e0ebab4b-62db-4b77-b730-0413b51bdeaf","Type":"ContainerDied","Data":"8fcfa123058cc29ae9f50e89a458a6e6809ba67a34f713324bd10fe0bf31b5b2"} Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.718012 4677 scope.go:117] "RemoveContainer" containerID="1d923eb357efa003508d4e717adee4aa5a17e8572cf26b077ee79b1c77ca8dd8" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.718013 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c98758bb9-jwrht" Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.723425 4677 generic.go:334] "Generic (PLEG): container finished" podID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerID="131b5669e7bdb8fec5a14c35b257c557a9085a95a7fe9a66bf08a18a6b349756" exitCode=0 Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.723484 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerDied","Data":"131b5669e7bdb8fec5a14c35b257c557a9085a95a7fe9a66bf08a18a6b349756"} Dec 03 14:09:25 crc kubenswrapper[4677]: I1203 14:09:25.739415 4677 scope.go:117] "RemoveContainer" containerID="ad1edec02b060afc9ba51b3a8a2d1a56e70d94e171d3c3dc8307896909ffc7f5" Dec 03 14:09:26 crc kubenswrapper[4677]: I1203 14:09:26.017249 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c98758bb9-jwrht"] Dec 03 14:09:26 crc kubenswrapper[4677]: I1203 14:09:26.017508 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c98758bb9-jwrht"] Dec 03 14:09:26 crc kubenswrapper[4677]: I1203 14:09:26.735633 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerStarted","Data":"f209203ed74dc0d31ade3ef18bd09b8f182817995dd2b9969d9eea2a59c4efd7"} Dec 03 14:09:27 crc kubenswrapper[4677]: I1203 14:09:27.325129 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:09:27 crc kubenswrapper[4677]: I1203 14:09:27.364152 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Dec 03 14:09:27 crc kubenswrapper[4677]: I1203 14:09:27.707166 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 14:09:27 crc kubenswrapper[4677]: I1203 14:09:27.986422 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" path="/var/lib/kubelet/pods/e0ebab4b-62db-4b77-b730-0413b51bdeaf/volumes" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.595186 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ncgkf"] Dec 03 14:09:28 crc kubenswrapper[4677]: E1203 14:09:28.595877 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerName="dnsmasq-dns" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.595901 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerName="dnsmasq-dns" Dec 03 14:09:28 crc kubenswrapper[4677]: E1203 14:09:28.595940 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerName="init" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.595966 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerName="init" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.596159 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ebab4b-62db-4b77-b730-0413b51bdeaf" containerName="dnsmasq-dns" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.596798 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.609571 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ncgkf"] Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.662187 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mxr2\" (UniqueName: \"kubernetes.io/projected/0942d3aa-be7a-47e1-a5a5-7fcae1993259-kube-api-access-6mxr2\") pod \"glance-db-create-ncgkf\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.662269 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0942d3aa-be7a-47e1-a5a5-7fcae1993259-operator-scripts\") pod \"glance-db-create-ncgkf\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.705201 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6be3-account-create-update-pj4dm"] Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.706290 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.708308 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.744664 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6be3-account-create-update-pj4dm"] Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.764284 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mxr2\" (UniqueName: \"kubernetes.io/projected/0942d3aa-be7a-47e1-a5a5-7fcae1993259-kube-api-access-6mxr2\") pod \"glance-db-create-ncgkf\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.767764 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds6c7\" (UniqueName: \"kubernetes.io/projected/a8853a43-b0b5-46da-932d-53826ee8cd4e-kube-api-access-ds6c7\") pod \"glance-6be3-account-create-update-pj4dm\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.767928 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0942d3aa-be7a-47e1-a5a5-7fcae1993259-operator-scripts\") pod \"glance-db-create-ncgkf\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.768073 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8853a43-b0b5-46da-932d-53826ee8cd4e-operator-scripts\") pod \"glance-6be3-account-create-update-pj4dm\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.769366 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0942d3aa-be7a-47e1-a5a5-7fcae1993259-operator-scripts\") pod \"glance-db-create-ncgkf\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.786475 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mxr2\" (UniqueName: \"kubernetes.io/projected/0942d3aa-be7a-47e1-a5a5-7fcae1993259-kube-api-access-6mxr2\") pod \"glance-db-create-ncgkf\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.869640 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds6c7\" (UniqueName: \"kubernetes.io/projected/a8853a43-b0b5-46da-932d-53826ee8cd4e-kube-api-access-ds6c7\") pod \"glance-6be3-account-create-update-pj4dm\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.870276 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8853a43-b0b5-46da-932d-53826ee8cd4e-operator-scripts\") pod \"glance-6be3-account-create-update-pj4dm\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.870906 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8853a43-b0b5-46da-932d-53826ee8cd4e-operator-scripts\") pod \"glance-6be3-account-create-update-pj4dm\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.887496 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds6c7\" (UniqueName: \"kubernetes.io/projected/a8853a43-b0b5-46da-932d-53826ee8cd4e-kube-api-access-ds6c7\") pod \"glance-6be3-account-create-update-pj4dm\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:28 crc kubenswrapper[4677]: I1203 14:09:28.926997 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.038122 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.213827 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ncgkf"] Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.542851 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6be3-account-create-update-pj4dm"] Dec 03 14:09:29 crc kubenswrapper[4677]: W1203 14:09:29.545922 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8853a43_b0b5_46da_932d_53826ee8cd4e.slice/crio-14008cd887b1bba81a5fec066008143dc207f3cf4f1f36dea277e14594f38abb WatchSource:0}: Error finding container 14008cd887b1bba81a5fec066008143dc207f3cf4f1f36dea277e14594f38abb: Status 404 returned error can't find the container with id 14008cd887b1bba81a5fec066008143dc207f3cf4f1f36dea277e14594f38abb Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.766727 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerStarted","Data":"2f216e09675889656977525ecea68681584e97f744c6a2d705d35a6612fa5d86"} Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.766768 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerStarted","Data":"f64a2560b1753ecb052a6d47e806c54eeff5b227bd7f96324fd2c656a66c9f84"} Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.770364 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6be3-account-create-update-pj4dm" event={"ID":"a8853a43-b0b5-46da-932d-53826ee8cd4e","Type":"ContainerStarted","Data":"9b976fe0d50cabd937b9fccfef33cda3f678185496529614cd2d77251bc04f72"} Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.770406 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6be3-account-create-update-pj4dm" event={"ID":"a8853a43-b0b5-46da-932d-53826ee8cd4e","Type":"ContainerStarted","Data":"14008cd887b1bba81a5fec066008143dc207f3cf4f1f36dea277e14594f38abb"} Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.771830 4677 generic.go:334] "Generic (PLEG): container finished" podID="0942d3aa-be7a-47e1-a5a5-7fcae1993259" containerID="31c5780adee53a71145c6e49c64f0976ad29c84c1f9dbdc46098f1666a07d86f" exitCode=0 Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.771876 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ncgkf" event={"ID":"0942d3aa-be7a-47e1-a5a5-7fcae1993259","Type":"ContainerDied","Data":"31c5780adee53a71145c6e49c64f0976ad29c84c1f9dbdc46098f1666a07d86f"} Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.771928 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ncgkf" event={"ID":"0942d3aa-be7a-47e1-a5a5-7fcae1993259","Type":"ContainerStarted","Data":"7dcadc67e3770e39f4879f4992f7d31682ceacfd80dfbb54fa5b882a4d0b1475"} Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.794675 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.794661001 podStartE2EDuration="16.794661001s" podCreationTimestamp="2025-12-03 14:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:29.792426395 +0000 UTC m=+1360.538758880" watchObservedRunningTime="2025-12-03 14:09:29.794661001 +0000 UTC m=+1360.540993456" Dec 03 14:09:29 crc kubenswrapper[4677]: I1203 14:09:29.841751 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-6be3-account-create-update-pj4dm" podStartSLOduration=1.841732125 podStartE2EDuration="1.841732125s" podCreationTimestamp="2025-12-03 14:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:29.837826156 +0000 UTC m=+1360.584158611" watchObservedRunningTime="2025-12-03 14:09:29.841732125 +0000 UTC m=+1360.588064580" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.230051 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-r2jg2"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.231377 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.245221 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-2f2f-account-create-update-wvblx"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.246353 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.248105 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.264852 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-r2jg2"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.296707 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbkjc\" (UniqueName: \"kubernetes.io/projected/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-kube-api-access-xbkjc\") pod \"cinder-db-create-r2jg2\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.296827 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-operator-scripts\") pod \"cinder-db-create-r2jg2\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.296874 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxqjc\" (UniqueName: \"kubernetes.io/projected/20bee223-5339-4fe5-a341-9bd1b1904079-kube-api-access-zxqjc\") pod \"cinder-2f2f-account-create-update-wvblx\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.296902 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20bee223-5339-4fe5-a341-9bd1b1904079-operator-scripts\") pod \"cinder-2f2f-account-create-update-wvblx\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.329785 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2f2f-account-create-update-wvblx"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.378308 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-cdnh4"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.379779 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.385552 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cdnh4"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.397946 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-operator-scripts\") pod \"cinder-db-create-r2jg2\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.398015 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx2nw\" (UniqueName: \"kubernetes.io/projected/ea30c0a1-61fd-45e9-905b-0f022a6381bd-kube-api-access-vx2nw\") pod \"barbican-db-create-cdnh4\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.398046 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxqjc\" (UniqueName: \"kubernetes.io/projected/20bee223-5339-4fe5-a341-9bd1b1904079-kube-api-access-zxqjc\") pod \"cinder-2f2f-account-create-update-wvblx\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.398079 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20bee223-5339-4fe5-a341-9bd1b1904079-operator-scripts\") pod \"cinder-2f2f-account-create-update-wvblx\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.398118 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea30c0a1-61fd-45e9-905b-0f022a6381bd-operator-scripts\") pod \"barbican-db-create-cdnh4\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.398146 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbkjc\" (UniqueName: \"kubernetes.io/projected/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-kube-api-access-xbkjc\") pod \"cinder-db-create-r2jg2\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.399171 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-operator-scripts\") pod \"cinder-db-create-r2jg2\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.399848 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20bee223-5339-4fe5-a341-9bd1b1904079-operator-scripts\") pod \"cinder-2f2f-account-create-update-wvblx\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.410781 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-94ppc"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.412056 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.415145 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.415468 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-f6mpt" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.422319 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbkjc\" (UniqueName: \"kubernetes.io/projected/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-kube-api-access-xbkjc\") pod \"cinder-db-create-r2jg2\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.427064 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-94ppc"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.431578 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxqjc\" (UniqueName: \"kubernetes.io/projected/20bee223-5339-4fe5-a341-9bd1b1904079-kube-api-access-zxqjc\") pod \"cinder-2f2f-account-create-update-wvblx\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.499374 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea30c0a1-61fd-45e9-905b-0f022a6381bd-operator-scripts\") pod \"barbican-db-create-cdnh4\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.500108 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-config-data\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.500045 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea30c0a1-61fd-45e9-905b-0f022a6381bd-operator-scripts\") pod \"barbican-db-create-cdnh4\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.500279 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-db-sync-config-data\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.500359 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-combined-ca-bundle\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.500395 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx2nw\" (UniqueName: \"kubernetes.io/projected/ea30c0a1-61fd-45e9-905b-0f022a6381bd-kube-api-access-vx2nw\") pod \"barbican-db-create-cdnh4\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.500466 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2nvc\" (UniqueName: \"kubernetes.io/projected/3aaf4f81-d89b-40d9-af15-4920cdf73a76-kube-api-access-s2nvc\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.523019 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6850-account-create-update-8lrt8"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.524276 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.524658 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx2nw\" (UniqueName: \"kubernetes.io/projected/ea30c0a1-61fd-45e9-905b-0f022a6381bd-kube-api-access-vx2nw\") pod \"barbican-db-create-cdnh4\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.528220 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.538325 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6850-account-create-update-8lrt8"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.561263 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-r9jq5"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.593889 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.609636 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610612 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-config-data\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610648 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2nvc\" (UniqueName: \"kubernetes.io/projected/3aaf4f81-d89b-40d9-af15-4920cdf73a76-kube-api-access-s2nvc\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610685 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7nww\" (UniqueName: \"kubernetes.io/projected/b91633d9-71f9-40bc-832d-b24143d8478b-kube-api-access-k7nww\") pod \"barbican-6850-account-create-update-8lrt8\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610727 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-config-data\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610771 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-combined-ca-bundle\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610798 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-db-sync-config-data\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610823 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7bkw\" (UniqueName: \"kubernetes.io/projected/711edb9c-bc87-4f0a-adea-36ef2502783f-kube-api-access-h7bkw\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610852 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91633d9-71f9-40bc-832d-b24143d8478b-operator-scripts\") pod \"barbican-6850-account-create-update-8lrt8\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.610888 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-combined-ca-bundle\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.617291 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-db-sync-config-data\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.621258 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.621268 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.621554 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s8xkt" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.624794 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.633469 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-config-data\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.635995 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.652697 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-combined-ca-bundle\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.653152 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2nvc\" (UniqueName: \"kubernetes.io/projected/3aaf4f81-d89b-40d9-af15-4920cdf73a76-kube-api-access-s2nvc\") pod \"watcher-db-sync-94ppc\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.663026 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-r9jq5"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.699131 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-94ctf"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.705617 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.712639 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7bkw\" (UniqueName: \"kubernetes.io/projected/711edb9c-bc87-4f0a-adea-36ef2502783f-kube-api-access-h7bkw\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.712686 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d278aed-ed7e-49fe-b952-899f9d0076ef-operator-scripts\") pod \"neutron-db-create-94ctf\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.712712 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91633d9-71f9-40bc-832d-b24143d8478b-operator-scripts\") pod \"barbican-6850-account-create-update-8lrt8\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.712777 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-config-data\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.712806 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7nww\" (UniqueName: \"kubernetes.io/projected/b91633d9-71f9-40bc-832d-b24143d8478b-kube-api-access-k7nww\") pod \"barbican-6850-account-create-update-8lrt8\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.712825 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6hbv\" (UniqueName: \"kubernetes.io/projected/7d278aed-ed7e-49fe-b952-899f9d0076ef-kube-api-access-m6hbv\") pod \"neutron-db-create-94ctf\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.712871 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-combined-ca-bundle\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.713454 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91633d9-71f9-40bc-832d-b24143d8478b-operator-scripts\") pod \"barbican-6850-account-create-update-8lrt8\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.714057 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.715025 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-94ctf"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.723073 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-combined-ca-bundle\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.741293 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7bkw\" (UniqueName: \"kubernetes.io/projected/711edb9c-bc87-4f0a-adea-36ef2502783f-kube-api-access-h7bkw\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.747903 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7nww\" (UniqueName: \"kubernetes.io/projected/b91633d9-71f9-40bc-832d-b24143d8478b-kube-api-access-k7nww\") pod \"barbican-6850-account-create-update-8lrt8\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.748653 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-config-data\") pod \"keystone-db-sync-r9jq5\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.814358 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d278aed-ed7e-49fe-b952-899f9d0076ef-operator-scripts\") pod \"neutron-db-create-94ctf\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.814494 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6hbv\" (UniqueName: \"kubernetes.io/projected/7d278aed-ed7e-49fe-b952-899f9d0076ef-kube-api-access-m6hbv\") pod \"neutron-db-create-94ctf\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.814969 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.815262 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d278aed-ed7e-49fe-b952-899f9d0076ef-operator-scripts\") pod \"neutron-db-create-94ctf\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.819895 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-cf68-account-create-update-wppvc"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.822220 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.827411 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.867575 4677 generic.go:334] "Generic (PLEG): container finished" podID="a8853a43-b0b5-46da-932d-53826ee8cd4e" containerID="9b976fe0d50cabd937b9fccfef33cda3f678185496529614cd2d77251bc04f72" exitCode=0 Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.867806 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6be3-account-create-update-pj4dm" event={"ID":"a8853a43-b0b5-46da-932d-53826ee8cd4e","Type":"ContainerDied","Data":"9b976fe0d50cabd937b9fccfef33cda3f678185496529614cd2d77251bc04f72"} Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.872209 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cf68-account-create-update-wppvc"] Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.876922 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6hbv\" (UniqueName: \"kubernetes.io/projected/7d278aed-ed7e-49fe-b952-899f9d0076ef-kube-api-access-m6hbv\") pod \"neutron-db-create-94ctf\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.916145 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d336d821-9f68-4486-95f7-e8c79c65a928-operator-scripts\") pod \"neutron-cf68-account-create-update-wppvc\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.916215 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr2zh\" (UniqueName: \"kubernetes.io/projected/d336d821-9f68-4486-95f7-e8c79c65a928-kube-api-access-zr2zh\") pod \"neutron-cf68-account-create-update-wppvc\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:30 crc kubenswrapper[4677]: I1203 14:09:30.992524 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.002218 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.018048 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d336d821-9f68-4486-95f7-e8c79c65a928-operator-scripts\") pod \"neutron-cf68-account-create-update-wppvc\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.018153 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr2zh\" (UniqueName: \"kubernetes.io/projected/d336d821-9f68-4486-95f7-e8c79c65a928-kube-api-access-zr2zh\") pod \"neutron-cf68-account-create-update-wppvc\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.032341 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d336d821-9f68-4486-95f7-e8c79c65a928-operator-scripts\") pod \"neutron-cf68-account-create-update-wppvc\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.039576 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr2zh\" (UniqueName: \"kubernetes.io/projected/d336d821-9f68-4486-95f7-e8c79c65a928-kube-api-access-zr2zh\") pod \"neutron-cf68-account-create-update-wppvc\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.053071 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.228164 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.249899 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2f2f-account-create-update-wvblx"] Dec 03 14:09:31 crc kubenswrapper[4677]: W1203 14:09:31.276140 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20bee223_5339_4fe5_a341_9bd1b1904079.slice/crio-bf45e3b2dac0fd7046935ba1266801879b507fbe0d50ed475071b1a394fc17b8 WatchSource:0}: Error finding container bf45e3b2dac0fd7046935ba1266801879b507fbe0d50ed475071b1a394fc17b8: Status 404 returned error can't find the container with id bf45e3b2dac0fd7046935ba1266801879b507fbe0d50ed475071b1a394fc17b8 Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.427623 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-r2jg2"] Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.432066 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.529813 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0942d3aa-be7a-47e1-a5a5-7fcae1993259-operator-scripts\") pod \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.529873 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mxr2\" (UniqueName: \"kubernetes.io/projected/0942d3aa-be7a-47e1-a5a5-7fcae1993259-kube-api-access-6mxr2\") pod \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\" (UID: \"0942d3aa-be7a-47e1-a5a5-7fcae1993259\") " Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.530769 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0942d3aa-be7a-47e1-a5a5-7fcae1993259-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0942d3aa-be7a-47e1-a5a5-7fcae1993259" (UID: "0942d3aa-be7a-47e1-a5a5-7fcae1993259"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.539897 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0942d3aa-be7a-47e1-a5a5-7fcae1993259-kube-api-access-6mxr2" (OuterVolumeSpecName: "kube-api-access-6mxr2") pod "0942d3aa-be7a-47e1-a5a5-7fcae1993259" (UID: "0942d3aa-be7a-47e1-a5a5-7fcae1993259"). InnerVolumeSpecName "kube-api-access-6mxr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:31 crc kubenswrapper[4677]: E1203 14:09:31.540309 4677 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.162:46274->38.102.83.162:39005: read tcp 38.102.83.162:46274->38.102.83.162:39005: read: connection reset by peer Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.605054 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-94ppc"] Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.626837 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cdnh4"] Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.633843 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0942d3aa-be7a-47e1-a5a5-7fcae1993259-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.633874 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mxr2\" (UniqueName: \"kubernetes.io/projected/0942d3aa-be7a-47e1-a5a5-7fcae1993259-kube-api-access-6mxr2\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.790089 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6850-account-create-update-8lrt8"] Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.804325 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-r9jq5"] Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.902269 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-94ppc" event={"ID":"3aaf4f81-d89b-40d9-af15-4920cdf73a76","Type":"ContainerStarted","Data":"2610564cf7ef4855e34ff1b21aec2b8f0152476df229d9c93957996cc1483130"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.907159 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-r2jg2" event={"ID":"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3","Type":"ContainerStarted","Data":"ff98e5e99d608250b4255d7926137dc9bb4b9476c5a6305d6e3a01940a37a6a2"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.907237 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-r2jg2" event={"ID":"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3","Type":"ContainerStarted","Data":"8f596467e25b69273f61756f42564d869bd99e37539b7d0755becc08faa715a3"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.913830 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-94ctf"] Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.925257 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f2f-account-create-update-wvblx" event={"ID":"20bee223-5339-4fe5-a341-9bd1b1904079","Type":"ContainerStarted","Data":"40f7a4ad7c5d67b5338d8fed11de0ff8501f0df2a42f5b4cefe0f762f8082d8c"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.925292 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f2f-account-create-update-wvblx" event={"ID":"20bee223-5339-4fe5-a341-9bd1b1904079","Type":"ContainerStarted","Data":"bf45e3b2dac0fd7046935ba1266801879b507fbe0d50ed475071b1a394fc17b8"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.942015 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-r2jg2" podStartSLOduration=1.941995555 podStartE2EDuration="1.941995555s" podCreationTimestamp="2025-12-03 14:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:31.938935317 +0000 UTC m=+1362.685267772" watchObservedRunningTime="2025-12-03 14:09:31.941995555 +0000 UTC m=+1362.688328020" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.943484 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ncgkf" event={"ID":"0942d3aa-be7a-47e1-a5a5-7fcae1993259","Type":"ContainerDied","Data":"7dcadc67e3770e39f4879f4992f7d31682ceacfd80dfbb54fa5b882a4d0b1475"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.943536 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dcadc67e3770e39f4879f4992f7d31682ceacfd80dfbb54fa5b882a4d0b1475" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.943499 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ncgkf" Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.946925 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6850-account-create-update-8lrt8" event={"ID":"b91633d9-71f9-40bc-832d-b24143d8478b","Type":"ContainerStarted","Data":"87d36cfd533d7f6452aca85ab8ab37c160d140a209326c291cbe75440805e0f5"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.951283 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r9jq5" event={"ID":"711edb9c-bc87-4f0a-adea-36ef2502783f","Type":"ContainerStarted","Data":"649c6f2ea7fa7d48f475675a497fcbde08d05d996982b7c9842daaa2161defea"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.954406 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cdnh4" event={"ID":"ea30c0a1-61fd-45e9-905b-0f022a6381bd","Type":"ContainerStarted","Data":"eeb3279669388080e59270473bb3fd077f6f32ab181915561302cd7fab2ddd61"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.954443 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cdnh4" event={"ID":"ea30c0a1-61fd-45e9-905b-0f022a6381bd","Type":"ContainerStarted","Data":"502142ba91e206ae1473acb47ab45214bf8313167c61594dff5d8236b00b9dfd"} Dec 03 14:09:31 crc kubenswrapper[4677]: I1203 14:09:31.993648 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-2f2f-account-create-update-wvblx" podStartSLOduration=1.993624945 podStartE2EDuration="1.993624945s" podCreationTimestamp="2025-12-03 14:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:31.98870263 +0000 UTC m=+1362.735035085" watchObservedRunningTime="2025-12-03 14:09:31.993624945 +0000 UTC m=+1362.739957400" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.011802 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-cdnh4" podStartSLOduration=2.01178069 podStartE2EDuration="2.01178069s" podCreationTimestamp="2025-12-03 14:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:32.003757864 +0000 UTC m=+1362.750090329" watchObservedRunningTime="2025-12-03 14:09:32.01178069 +0000 UTC m=+1362.758113145" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.095752 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cf68-account-create-update-wppvc"] Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.298778 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.456073 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds6c7\" (UniqueName: \"kubernetes.io/projected/a8853a43-b0b5-46da-932d-53826ee8cd4e-kube-api-access-ds6c7\") pod \"a8853a43-b0b5-46da-932d-53826ee8cd4e\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.456268 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8853a43-b0b5-46da-932d-53826ee8cd4e-operator-scripts\") pod \"a8853a43-b0b5-46da-932d-53826ee8cd4e\" (UID: \"a8853a43-b0b5-46da-932d-53826ee8cd4e\") " Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.456885 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8853a43-b0b5-46da-932d-53826ee8cd4e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8853a43-b0b5-46da-932d-53826ee8cd4e" (UID: "a8853a43-b0b5-46da-932d-53826ee8cd4e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.457278 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8853a43-b0b5-46da-932d-53826ee8cd4e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.465222 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8853a43-b0b5-46da-932d-53826ee8cd4e-kube-api-access-ds6c7" (OuterVolumeSpecName: "kube-api-access-ds6c7") pod "a8853a43-b0b5-46da-932d-53826ee8cd4e" (UID: "a8853a43-b0b5-46da-932d-53826ee8cd4e"). InnerVolumeSpecName "kube-api-access-ds6c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.559219 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds6c7\" (UniqueName: \"kubernetes.io/projected/a8853a43-b0b5-46da-932d-53826ee8cd4e-kube-api-access-ds6c7\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.962916 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6be3-account-create-update-pj4dm" event={"ID":"a8853a43-b0b5-46da-932d-53826ee8cd4e","Type":"ContainerDied","Data":"14008cd887b1bba81a5fec066008143dc207f3cf4f1f36dea277e14594f38abb"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.963270 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14008cd887b1bba81a5fec066008143dc207f3cf4f1f36dea277e14594f38abb" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.963325 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6be3-account-create-update-pj4dm" Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.970829 4677 generic.go:334] "Generic (PLEG): container finished" podID="7d278aed-ed7e-49fe-b952-899f9d0076ef" containerID="6bb0a2c1588e6f8de958dc087c06cef727b1c065aca436183fd2cda4fdd92192" exitCode=0 Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.970896 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-94ctf" event={"ID":"7d278aed-ed7e-49fe-b952-899f9d0076ef","Type":"ContainerDied","Data":"6bb0a2c1588e6f8de958dc087c06cef727b1c065aca436183fd2cda4fdd92192"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.971199 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-94ctf" event={"ID":"7d278aed-ed7e-49fe-b952-899f9d0076ef","Type":"ContainerStarted","Data":"b2792506ad18c839f8e5b1c5af1bc9dc9cbb2bc155924ff4746dfaf8f1b19f1e"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.974888 4677 generic.go:334] "Generic (PLEG): container finished" podID="b91633d9-71f9-40bc-832d-b24143d8478b" containerID="49cc87b587f88610d2dd11b90c14ba03581c6f96ecdc90e9aaca9eb7d6623e48" exitCode=0 Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.974937 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6850-account-create-update-8lrt8" event={"ID":"b91633d9-71f9-40bc-832d-b24143d8478b","Type":"ContainerDied","Data":"49cc87b587f88610d2dd11b90c14ba03581c6f96ecdc90e9aaca9eb7d6623e48"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.977644 4677 generic.go:334] "Generic (PLEG): container finished" podID="ea30c0a1-61fd-45e9-905b-0f022a6381bd" containerID="eeb3279669388080e59270473bb3fd077f6f32ab181915561302cd7fab2ddd61" exitCode=0 Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.977703 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cdnh4" event={"ID":"ea30c0a1-61fd-45e9-905b-0f022a6381bd","Type":"ContainerDied","Data":"eeb3279669388080e59270473bb3fd077f6f32ab181915561302cd7fab2ddd61"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.979938 4677 generic.go:334] "Generic (PLEG): container finished" podID="d336d821-9f68-4486-95f7-e8c79c65a928" containerID="eddaa4406812d8cc571e839129858025034f68c41b6c802b490846d25e716fb3" exitCode=0 Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.979986 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cf68-account-create-update-wppvc" event={"ID":"d336d821-9f68-4486-95f7-e8c79c65a928","Type":"ContainerDied","Data":"eddaa4406812d8cc571e839129858025034f68c41b6c802b490846d25e716fb3"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.980006 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cf68-account-create-update-wppvc" event={"ID":"d336d821-9f68-4486-95f7-e8c79c65a928","Type":"ContainerStarted","Data":"0a2852f06fd00baa2828187789207b1139d90061c2ad4595e990f521049858b2"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.987675 4677 generic.go:334] "Generic (PLEG): container finished" podID="dded6789-7bf2-4fd7-8d73-6a11c4ad48d3" containerID="ff98e5e99d608250b4255d7926137dc9bb4b9476c5a6305d6e3a01940a37a6a2" exitCode=0 Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.987740 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-r2jg2" event={"ID":"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3","Type":"ContainerDied","Data":"ff98e5e99d608250b4255d7926137dc9bb4b9476c5a6305d6e3a01940a37a6a2"} Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.994014 4677 generic.go:334] "Generic (PLEG): container finished" podID="20bee223-5339-4fe5-a341-9bd1b1904079" containerID="40f7a4ad7c5d67b5338d8fed11de0ff8501f0df2a42f5b4cefe0f762f8082d8c" exitCode=0 Dec 03 14:09:32 crc kubenswrapper[4677]: I1203 14:09:32.994060 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f2f-account-create-update-wvblx" event={"ID":"20bee223-5339-4fe5-a341-9bd1b1904079","Type":"ContainerDied","Data":"40f7a4ad7c5d67b5338d8fed11de0ff8501f0df2a42f5b4cefe0f762f8082d8c"} Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.846268 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-kbkxk"] Dec 03 14:09:33 crc kubenswrapper[4677]: E1203 14:09:33.846808 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0942d3aa-be7a-47e1-a5a5-7fcae1993259" containerName="mariadb-database-create" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.846836 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0942d3aa-be7a-47e1-a5a5-7fcae1993259" containerName="mariadb-database-create" Dec 03 14:09:33 crc kubenswrapper[4677]: E1203 14:09:33.846896 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8853a43-b0b5-46da-932d-53826ee8cd4e" containerName="mariadb-account-create-update" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.846909 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8853a43-b0b5-46da-932d-53826ee8cd4e" containerName="mariadb-account-create-update" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.847212 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0942d3aa-be7a-47e1-a5a5-7fcae1993259" containerName="mariadb-database-create" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.847250 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8853a43-b0b5-46da-932d-53826ee8cd4e" containerName="mariadb-account-create-update" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.848072 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.850552 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dl8zp" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.850720 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.857127 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kbkxk"] Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.994369 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-db-sync-config-data\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.994610 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-combined-ca-bundle\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.995359 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq6p8\" (UniqueName: \"kubernetes.io/projected/4c12c954-d94a-404a-9263-a8371fdc12e4-kube-api-access-bq6p8\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:33 crc kubenswrapper[4677]: I1203 14:09:33.995444 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-config-data\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.097204 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq6p8\" (UniqueName: \"kubernetes.io/projected/4c12c954-d94a-404a-9263-a8371fdc12e4-kube-api-access-bq6p8\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.097461 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-config-data\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.097493 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-db-sync-config-data\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.097532 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-combined-ca-bundle\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.106087 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-combined-ca-bundle\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.107356 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-db-sync-config-data\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.119378 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-config-data\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.134078 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq6p8\" (UniqueName: \"kubernetes.io/projected/4c12c954-d94a-404a-9263-a8371fdc12e4-kube-api-access-bq6p8\") pod \"glance-db-sync-kbkxk\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.177838 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbkxk" Dec 03 14:09:34 crc kubenswrapper[4677]: I1203 14:09:34.343178 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.475807 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.482249 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.488533 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.499811 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.518529 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643134 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr2zh\" (UniqueName: \"kubernetes.io/projected/d336d821-9f68-4486-95f7-e8c79c65a928-kube-api-access-zr2zh\") pod \"d336d821-9f68-4486-95f7-e8c79c65a928\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643212 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-operator-scripts\") pod \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643303 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbkjc\" (UniqueName: \"kubernetes.io/projected/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-kube-api-access-xbkjc\") pod \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\" (UID: \"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643331 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx2nw\" (UniqueName: \"kubernetes.io/projected/ea30c0a1-61fd-45e9-905b-0f022a6381bd-kube-api-access-vx2nw\") pod \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643376 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d336d821-9f68-4486-95f7-e8c79c65a928-operator-scripts\") pod \"d336d821-9f68-4486-95f7-e8c79c65a928\" (UID: \"d336d821-9f68-4486-95f7-e8c79c65a928\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643402 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20bee223-5339-4fe5-a341-9bd1b1904079-operator-scripts\") pod \"20bee223-5339-4fe5-a341-9bd1b1904079\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643427 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea30c0a1-61fd-45e9-905b-0f022a6381bd-operator-scripts\") pod \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\" (UID: \"ea30c0a1-61fd-45e9-905b-0f022a6381bd\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643448 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxqjc\" (UniqueName: \"kubernetes.io/projected/20bee223-5339-4fe5-a341-9bd1b1904079-kube-api-access-zxqjc\") pod \"20bee223-5339-4fe5-a341-9bd1b1904079\" (UID: \"20bee223-5339-4fe5-a341-9bd1b1904079\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643463 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d278aed-ed7e-49fe-b952-899f9d0076ef-operator-scripts\") pod \"7d278aed-ed7e-49fe-b952-899f9d0076ef\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.643510 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6hbv\" (UniqueName: \"kubernetes.io/projected/7d278aed-ed7e-49fe-b952-899f9d0076ef-kube-api-access-m6hbv\") pod \"7d278aed-ed7e-49fe-b952-899f9d0076ef\" (UID: \"7d278aed-ed7e-49fe-b952-899f9d0076ef\") " Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.644122 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20bee223-5339-4fe5-a341-9bd1b1904079-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20bee223-5339-4fe5-a341-9bd1b1904079" (UID: "20bee223-5339-4fe5-a341-9bd1b1904079"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.644268 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea30c0a1-61fd-45e9-905b-0f022a6381bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea30c0a1-61fd-45e9-905b-0f022a6381bd" (UID: "ea30c0a1-61fd-45e9-905b-0f022a6381bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.644339 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d278aed-ed7e-49fe-b952-899f9d0076ef-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7d278aed-ed7e-49fe-b952-899f9d0076ef" (UID: "7d278aed-ed7e-49fe-b952-899f9d0076ef"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.644646 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d336d821-9f68-4486-95f7-e8c79c65a928-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d336d821-9f68-4486-95f7-e8c79c65a928" (UID: "d336d821-9f68-4486-95f7-e8c79c65a928"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.644904 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dded6789-7bf2-4fd7-8d73-6a11c4ad48d3" (UID: "dded6789-7bf2-4fd7-8d73-6a11c4ad48d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.649487 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20bee223-5339-4fe5-a341-9bd1b1904079-kube-api-access-zxqjc" (OuterVolumeSpecName: "kube-api-access-zxqjc") pod "20bee223-5339-4fe5-a341-9bd1b1904079" (UID: "20bee223-5339-4fe5-a341-9bd1b1904079"). InnerVolumeSpecName "kube-api-access-zxqjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.649523 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d336d821-9f68-4486-95f7-e8c79c65a928-kube-api-access-zr2zh" (OuterVolumeSpecName: "kube-api-access-zr2zh") pod "d336d821-9f68-4486-95f7-e8c79c65a928" (UID: "d336d821-9f68-4486-95f7-e8c79c65a928"). InnerVolumeSpecName "kube-api-access-zr2zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.649779 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea30c0a1-61fd-45e9-905b-0f022a6381bd-kube-api-access-vx2nw" (OuterVolumeSpecName: "kube-api-access-vx2nw") pod "ea30c0a1-61fd-45e9-905b-0f022a6381bd" (UID: "ea30c0a1-61fd-45e9-905b-0f022a6381bd"). InnerVolumeSpecName "kube-api-access-vx2nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.650131 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-kube-api-access-xbkjc" (OuterVolumeSpecName: "kube-api-access-xbkjc") pod "dded6789-7bf2-4fd7-8d73-6a11c4ad48d3" (UID: "dded6789-7bf2-4fd7-8d73-6a11c4ad48d3"). InnerVolumeSpecName "kube-api-access-xbkjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.664079 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d278aed-ed7e-49fe-b952-899f9d0076ef-kube-api-access-m6hbv" (OuterVolumeSpecName: "kube-api-access-m6hbv") pod "7d278aed-ed7e-49fe-b952-899f9d0076ef" (UID: "7d278aed-ed7e-49fe-b952-899f9d0076ef"). InnerVolumeSpecName "kube-api-access-m6hbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745392 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx2nw\" (UniqueName: \"kubernetes.io/projected/ea30c0a1-61fd-45e9-905b-0f022a6381bd-kube-api-access-vx2nw\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745445 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d336d821-9f68-4486-95f7-e8c79c65a928-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745463 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20bee223-5339-4fe5-a341-9bd1b1904079-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745477 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea30c0a1-61fd-45e9-905b-0f022a6381bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745491 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxqjc\" (UniqueName: \"kubernetes.io/projected/20bee223-5339-4fe5-a341-9bd1b1904079-kube-api-access-zxqjc\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745506 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d278aed-ed7e-49fe-b952-899f9d0076ef-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745525 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6hbv\" (UniqueName: \"kubernetes.io/projected/7d278aed-ed7e-49fe-b952-899f9d0076ef-kube-api-access-m6hbv\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745540 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr2zh\" (UniqueName: \"kubernetes.io/projected/d336d821-9f68-4486-95f7-e8c79c65a928-kube-api-access-zr2zh\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745557 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:36 crc kubenswrapper[4677]: I1203 14:09:36.745570 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbkjc\" (UniqueName: \"kubernetes.io/projected/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3-kube-api-access-xbkjc\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.039426 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cdnh4" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.039430 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cdnh4" event={"ID":"ea30c0a1-61fd-45e9-905b-0f022a6381bd","Type":"ContainerDied","Data":"502142ba91e206ae1473acb47ab45214bf8313167c61594dff5d8236b00b9dfd"} Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.039481 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="502142ba91e206ae1473acb47ab45214bf8313167c61594dff5d8236b00b9dfd" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.041270 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cf68-account-create-update-wppvc" event={"ID":"d336d821-9f68-4486-95f7-e8c79c65a928","Type":"ContainerDied","Data":"0a2852f06fd00baa2828187789207b1139d90061c2ad4595e990f521049858b2"} Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.041305 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a2852f06fd00baa2828187789207b1139d90061c2ad4595e990f521049858b2" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.041307 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cf68-account-create-update-wppvc" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.043351 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-r2jg2" event={"ID":"dded6789-7bf2-4fd7-8d73-6a11c4ad48d3","Type":"ContainerDied","Data":"8f596467e25b69273f61756f42564d869bd99e37539b7d0755becc08faa715a3"} Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.043374 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f596467e25b69273f61756f42564d869bd99e37539b7d0755becc08faa715a3" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.043432 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-r2jg2" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.046141 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f2f-account-create-update-wvblx" event={"ID":"20bee223-5339-4fe5-a341-9bd1b1904079","Type":"ContainerDied","Data":"bf45e3b2dac0fd7046935ba1266801879b507fbe0d50ed475071b1a394fc17b8"} Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.046191 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f2f-account-create-update-wvblx" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.046192 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf45e3b2dac0fd7046935ba1266801879b507fbe0d50ed475071b1a394fc17b8" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.048018 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-94ctf" event={"ID":"7d278aed-ed7e-49fe-b952-899f9d0076ef","Type":"ContainerDied","Data":"b2792506ad18c839f8e5b1c5af1bc9dc9cbb2bc155924ff4746dfaf8f1b19f1e"} Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.048051 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-94ctf" Dec 03 14:09:37 crc kubenswrapper[4677]: I1203 14:09:37.048061 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2792506ad18c839f8e5b1c5af1bc9dc9cbb2bc155924ff4746dfaf8f1b19f1e" Dec 03 14:09:38 crc kubenswrapper[4677]: I1203 14:09:38.437339 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:09:38 crc kubenswrapper[4677]: I1203 14:09:38.437648 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.104408 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6850-account-create-update-8lrt8" event={"ID":"b91633d9-71f9-40bc-832d-b24143d8478b","Type":"ContainerDied","Data":"87d36cfd533d7f6452aca85ab8ab37c160d140a209326c291cbe75440805e0f5"} Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.105005 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87d36cfd533d7f6452aca85ab8ab37c160d140a209326c291cbe75440805e0f5" Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.171394 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.355331 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7nww\" (UniqueName: \"kubernetes.io/projected/b91633d9-71f9-40bc-832d-b24143d8478b-kube-api-access-k7nww\") pod \"b91633d9-71f9-40bc-832d-b24143d8478b\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.355794 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91633d9-71f9-40bc-832d-b24143d8478b-operator-scripts\") pod \"b91633d9-71f9-40bc-832d-b24143d8478b\" (UID: \"b91633d9-71f9-40bc-832d-b24143d8478b\") " Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.356480 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91633d9-71f9-40bc-832d-b24143d8478b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b91633d9-71f9-40bc-832d-b24143d8478b" (UID: "b91633d9-71f9-40bc-832d-b24143d8478b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.362035 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b91633d9-71f9-40bc-832d-b24143d8478b-kube-api-access-k7nww" (OuterVolumeSpecName: "kube-api-access-k7nww") pod "b91633d9-71f9-40bc-832d-b24143d8478b" (UID: "b91633d9-71f9-40bc-832d-b24143d8478b"). InnerVolumeSpecName "kube-api-access-k7nww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.457632 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91633d9-71f9-40bc-832d-b24143d8478b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:42 crc kubenswrapper[4677]: I1203 14:09:42.457676 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7nww\" (UniqueName: \"kubernetes.io/projected/b91633d9-71f9-40bc-832d-b24143d8478b-kube-api-access-k7nww\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:43 crc kubenswrapper[4677]: I1203 14:09:43.115168 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r9jq5" event={"ID":"711edb9c-bc87-4f0a-adea-36ef2502783f","Type":"ContainerStarted","Data":"4b88e476aaf26876be26a3ce964bed0b3f0751c4207052c1d5986fbfdc7b3433"} Dec 03 14:09:43 crc kubenswrapper[4677]: I1203 14:09:43.127644 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6850-account-create-update-8lrt8" Dec 03 14:09:43 crc kubenswrapper[4677]: I1203 14:09:43.129879 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-94ppc" event={"ID":"3aaf4f81-d89b-40d9-af15-4920cdf73a76","Type":"ContainerStarted","Data":"b55fe36f06bde617e131cb67859b8773719c686e8ea20b017582dd9a095de032"} Dec 03 14:09:43 crc kubenswrapper[4677]: I1203 14:09:43.139312 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-r9jq5" podStartSLOduration=2.34254361 podStartE2EDuration="13.139295473s" podCreationTimestamp="2025-12-03 14:09:30 +0000 UTC" firstStartedPulling="2025-12-03 14:09:31.83507072 +0000 UTC m=+1362.581403175" lastFinishedPulling="2025-12-03 14:09:42.631822583 +0000 UTC m=+1373.378155038" observedRunningTime="2025-12-03 14:09:43.135396074 +0000 UTC m=+1373.881728529" watchObservedRunningTime="2025-12-03 14:09:43.139295473 +0000 UTC m=+1373.885627928" Dec 03 14:09:43 crc kubenswrapper[4677]: I1203 14:09:43.163513 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-94ppc" podStartSLOduration=2.074305789 podStartE2EDuration="13.163491462s" podCreationTimestamp="2025-12-03 14:09:30 +0000 UTC" firstStartedPulling="2025-12-03 14:09:31.619129157 +0000 UTC m=+1362.365461622" lastFinishedPulling="2025-12-03 14:09:42.70831484 +0000 UTC m=+1373.454647295" observedRunningTime="2025-12-03 14:09:43.158690339 +0000 UTC m=+1373.905022794" watchObservedRunningTime="2025-12-03 14:09:43.163491462 +0000 UTC m=+1373.909823917" Dec 03 14:09:43 crc kubenswrapper[4677]: I1203 14:09:43.239591 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kbkxk"] Dec 03 14:09:44 crc kubenswrapper[4677]: I1203 14:09:44.136471 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbkxk" event={"ID":"4c12c954-d94a-404a-9263-a8371fdc12e4","Type":"ContainerStarted","Data":"5f3f90c975a1130519cc70bf5fd2fe333ea91e0b04b54201e03255352eb01966"} Dec 03 14:09:44 crc kubenswrapper[4677]: I1203 14:09:44.343964 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:44 crc kubenswrapper[4677]: I1203 14:09:44.350794 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:45 crc kubenswrapper[4677]: I1203 14:09:45.152821 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 14:09:47 crc kubenswrapper[4677]: I1203 14:09:47.168855 4677 generic.go:334] "Generic (PLEG): container finished" podID="711edb9c-bc87-4f0a-adea-36ef2502783f" containerID="4b88e476aaf26876be26a3ce964bed0b3f0751c4207052c1d5986fbfdc7b3433" exitCode=0 Dec 03 14:09:47 crc kubenswrapper[4677]: I1203 14:09:47.168938 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r9jq5" event={"ID":"711edb9c-bc87-4f0a-adea-36ef2502783f","Type":"ContainerDied","Data":"4b88e476aaf26876be26a3ce964bed0b3f0751c4207052c1d5986fbfdc7b3433"} Dec 03 14:09:47 crc kubenswrapper[4677]: I1203 14:09:47.172188 4677 generic.go:334] "Generic (PLEG): container finished" podID="3aaf4f81-d89b-40d9-af15-4920cdf73a76" containerID="b55fe36f06bde617e131cb67859b8773719c686e8ea20b017582dd9a095de032" exitCode=0 Dec 03 14:09:47 crc kubenswrapper[4677]: I1203 14:09:47.172228 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-94ppc" event={"ID":"3aaf4f81-d89b-40d9-af15-4920cdf73a76","Type":"ContainerDied","Data":"b55fe36f06bde617e131cb67859b8773719c686e8ea20b017582dd9a095de032"} Dec 03 14:09:54 crc kubenswrapper[4677]: I1203 14:09:54.978678 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:54 crc kubenswrapper[4677]: I1203 14:09:54.986162 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.030832 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-config-data\") pod \"711edb9c-bc87-4f0a-adea-36ef2502783f\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.031416 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2nvc\" (UniqueName: \"kubernetes.io/projected/3aaf4f81-d89b-40d9-af15-4920cdf73a76-kube-api-access-s2nvc\") pod \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.031460 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-combined-ca-bundle\") pod \"711edb9c-bc87-4f0a-adea-36ef2502783f\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.031609 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-combined-ca-bundle\") pod \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.031726 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7bkw\" (UniqueName: \"kubernetes.io/projected/711edb9c-bc87-4f0a-adea-36ef2502783f-kube-api-access-h7bkw\") pod \"711edb9c-bc87-4f0a-adea-36ef2502783f\" (UID: \"711edb9c-bc87-4f0a-adea-36ef2502783f\") " Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.031762 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-config-data\") pod \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.031823 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-db-sync-config-data\") pod \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\" (UID: \"3aaf4f81-d89b-40d9-af15-4920cdf73a76\") " Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.037122 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3aaf4f81-d89b-40d9-af15-4920cdf73a76" (UID: "3aaf4f81-d89b-40d9-af15-4920cdf73a76"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.037244 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711edb9c-bc87-4f0a-adea-36ef2502783f-kube-api-access-h7bkw" (OuterVolumeSpecName: "kube-api-access-h7bkw") pod "711edb9c-bc87-4f0a-adea-36ef2502783f" (UID: "711edb9c-bc87-4f0a-adea-36ef2502783f"). InnerVolumeSpecName "kube-api-access-h7bkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.043059 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aaf4f81-d89b-40d9-af15-4920cdf73a76-kube-api-access-s2nvc" (OuterVolumeSpecName: "kube-api-access-s2nvc") pod "3aaf4f81-d89b-40d9-af15-4920cdf73a76" (UID: "3aaf4f81-d89b-40d9-af15-4920cdf73a76"). InnerVolumeSpecName "kube-api-access-s2nvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.060791 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "711edb9c-bc87-4f0a-adea-36ef2502783f" (UID: "711edb9c-bc87-4f0a-adea-36ef2502783f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.070707 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3aaf4f81-d89b-40d9-af15-4920cdf73a76" (UID: "3aaf4f81-d89b-40d9-af15-4920cdf73a76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.079917 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-config-data" (OuterVolumeSpecName: "config-data") pod "3aaf4f81-d89b-40d9-af15-4920cdf73a76" (UID: "3aaf4f81-d89b-40d9-af15-4920cdf73a76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.093248 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-config-data" (OuterVolumeSpecName: "config-data") pod "711edb9c-bc87-4f0a-adea-36ef2502783f" (UID: "711edb9c-bc87-4f0a-adea-36ef2502783f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.133972 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2nvc\" (UniqueName: \"kubernetes.io/projected/3aaf4f81-d89b-40d9-af15-4920cdf73a76-kube-api-access-s2nvc\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.134017 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.134032 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.134045 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7bkw\" (UniqueName: \"kubernetes.io/projected/711edb9c-bc87-4f0a-adea-36ef2502783f-kube-api-access-h7bkw\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.134057 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.134069 4677 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3aaf4f81-d89b-40d9-af15-4920cdf73a76-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.134082 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/711edb9c-bc87-4f0a-adea-36ef2502783f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.240087 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-94ppc" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.240080 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-94ppc" event={"ID":"3aaf4f81-d89b-40d9-af15-4920cdf73a76","Type":"ContainerDied","Data":"2610564cf7ef4855e34ff1b21aec2b8f0152476df229d9c93957996cc1483130"} Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.240223 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2610564cf7ef4855e34ff1b21aec2b8f0152476df229d9c93957996cc1483130" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.241492 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-r9jq5" event={"ID":"711edb9c-bc87-4f0a-adea-36ef2502783f","Type":"ContainerDied","Data":"649c6f2ea7fa7d48f475675a497fcbde08d05d996982b7c9842daaa2161defea"} Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.241517 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="649c6f2ea7fa7d48f475675a497fcbde08d05d996982b7c9842daaa2161defea" Dec 03 14:09:55 crc kubenswrapper[4677]: I1203 14:09:55.241565 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-r9jq5" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.309672 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8r7lf"] Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310330 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20bee223-5339-4fe5-a341-9bd1b1904079" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310344 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="20bee223-5339-4fe5-a341-9bd1b1904079" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310356 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d278aed-ed7e-49fe-b952-899f9d0076ef" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310362 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d278aed-ed7e-49fe-b952-899f9d0076ef" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310377 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d336d821-9f68-4486-95f7-e8c79c65a928" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310383 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="d336d821-9f68-4486-95f7-e8c79c65a928" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310396 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aaf4f81-d89b-40d9-af15-4920cdf73a76" containerName="watcher-db-sync" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310402 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aaf4f81-d89b-40d9-af15-4920cdf73a76" containerName="watcher-db-sync" Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310409 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dded6789-7bf2-4fd7-8d73-6a11c4ad48d3" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310414 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="dded6789-7bf2-4fd7-8d73-6a11c4ad48d3" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310426 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711edb9c-bc87-4f0a-adea-36ef2502783f" containerName="keystone-db-sync" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310431 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="711edb9c-bc87-4f0a-adea-36ef2502783f" containerName="keystone-db-sync" Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310444 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea30c0a1-61fd-45e9-905b-0f022a6381bd" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310449 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea30c0a1-61fd-45e9-905b-0f022a6381bd" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: E1203 14:09:56.310458 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b91633d9-71f9-40bc-832d-b24143d8478b" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310464 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b91633d9-71f9-40bc-832d-b24143d8478b" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310615 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea30c0a1-61fd-45e9-905b-0f022a6381bd" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310625 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="dded6789-7bf2-4fd7-8d73-6a11c4ad48d3" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310637 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="711edb9c-bc87-4f0a-adea-36ef2502783f" containerName="keystone-db-sync" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310648 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="20bee223-5339-4fe5-a341-9bd1b1904079" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310659 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="d336d821-9f68-4486-95f7-e8c79c65a928" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310670 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aaf4f81-d89b-40d9-af15-4920cdf73a76" containerName="watcher-db-sync" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310680 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b91633d9-71f9-40bc-832d-b24143d8478b" containerName="mariadb-account-create-update" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.310690 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d278aed-ed7e-49fe-b952-899f9d0076ef" containerName="mariadb-database-create" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.311274 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.312495 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbkxk" event={"ID":"4c12c954-d94a-404a-9263-a8371fdc12e4","Type":"ContainerStarted","Data":"8c5473373d1e776782a3579f5497a7aee84971d76b30289c6cc605d85f68a020"} Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.319378 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.323496 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.326264 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.326589 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.326759 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s8xkt" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.341056 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64cbd68bf9-nhcgh"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.342686 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.355289 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8r7lf"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363118 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf54x\" (UniqueName: \"kubernetes.io/projected/199679f6-a558-44e5-bd68-cc19b27a2f85-kube-api-access-wf54x\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363161 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-config-data\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363214 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-svc\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363229 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-config\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363276 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-nb\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363344 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-fernet-keys\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363367 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-sb\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363386 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-combined-ca-bundle\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363405 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-scripts\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363425 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65h8c\" (UniqueName: \"kubernetes.io/projected/d64c9f49-53ee-4069-af26-e2e01ead2ff3-kube-api-access-65h8c\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363440 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-credential-keys\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.363456 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-swift-storage-0\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.372003 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64cbd68bf9-nhcgh"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.465875 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-fernet-keys\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.465921 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-sb\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.465979 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-combined-ca-bundle\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466001 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-scripts\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466030 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-credential-keys\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466046 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65h8c\" (UniqueName: \"kubernetes.io/projected/d64c9f49-53ee-4069-af26-e2e01ead2ff3-kube-api-access-65h8c\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466063 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-swift-storage-0\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466087 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf54x\" (UniqueName: \"kubernetes.io/projected/199679f6-a558-44e5-bd68-cc19b27a2f85-kube-api-access-wf54x\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466109 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-config-data\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466155 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-svc\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466176 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-config\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.466208 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-nb\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.467410 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-nb\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.470940 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-sb\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.487192 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.488939 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.495012 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-svc\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.495511 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-config\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.496034 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-fernet-keys\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.496475 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-combined-ca-bundle\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.497395 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-scripts\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.503008 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-swift-storage-0\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.503534 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-f6mpt" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.503705 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.506548 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65h8c\" (UniqueName: \"kubernetes.io/projected/d64c9f49-53ee-4069-af26-e2e01ead2ff3-kube-api-access-65h8c\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.506871 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-credential-keys\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.509720 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-config-data\") pod \"keystone-bootstrap-8r7lf\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.522116 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.539528 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-kbkxk" podStartSLOduration=11.925225623 podStartE2EDuration="23.539509786s" podCreationTimestamp="2025-12-03 14:09:33 +0000 UTC" firstStartedPulling="2025-12-03 14:09:43.236661474 +0000 UTC m=+1373.982993939" lastFinishedPulling="2025-12-03 14:09:54.850945647 +0000 UTC m=+1385.597278102" observedRunningTime="2025-12-03 14:09:56.532278761 +0000 UTC m=+1387.278611256" watchObservedRunningTime="2025-12-03 14:09:56.539509786 +0000 UTC m=+1387.285842241" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.558303 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf54x\" (UniqueName: \"kubernetes.io/projected/199679f6-a558-44e5-bd68-cc19b27a2f85-kube-api-access-wf54x\") pod \"dnsmasq-dns-64cbd68bf9-nhcgh\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.572087 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf8zn\" (UniqueName: \"kubernetes.io/projected/5b533531-636f-4285-9c80-996c036eb4b1-kube-api-access-jf8zn\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.572152 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b533531-636f-4285-9c80-996c036eb4b1-logs\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.572199 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.572223 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-config-data\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.572245 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.611538 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-b7fdb6f4c-7tm5b"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.643151 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.662042 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.662702 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.662869 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-mp7tp" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.664139 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.668268 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.673729 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.675270 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.675887 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.678854 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679483 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-config-data\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679562 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679695 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thq65\" (UniqueName: \"kubernetes.io/projected/852927ce-1113-484c-a27f-5e9f75c7f7f0-kube-api-access-thq65\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679747 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/852927ce-1113-484c-a27f-5e9f75c7f7f0-horizon-secret-key\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679792 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852927ce-1113-484c-a27f-5e9f75c7f7f0-logs\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679821 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-scripts\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679852 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf8zn\" (UniqueName: \"kubernetes.io/projected/5b533531-636f-4285-9c80-996c036eb4b1-kube-api-access-jf8zn\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679926 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-config-data\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.679977 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b533531-636f-4285-9c80-996c036eb4b1-logs\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.680564 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b533531-636f-4285-9c80-996c036eb4b1-logs\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.695282 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.704135 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.707839 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.727267 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-config-data\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.751047 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf8zn\" (UniqueName: \"kubernetes.io/projected/5b533531-636f-4285-9c80-996c036eb4b1-kube-api-access-jf8zn\") pod \"watcher-api-0\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.762813 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b7fdb6f4c-7tm5b"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781192 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-config-data\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781238 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5edce1f1-9c11-4049-b286-a16939acb2e7-logs\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781265 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-config-data\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781337 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781377 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thq65\" (UniqueName: \"kubernetes.io/projected/852927ce-1113-484c-a27f-5e9f75c7f7f0-kube-api-access-thq65\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781397 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/852927ce-1113-484c-a27f-5e9f75c7f7f0-horizon-secret-key\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781420 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852927ce-1113-484c-a27f-5e9f75c7f7f0-logs\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781437 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-scripts\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.781459 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvqh6\" (UniqueName: \"kubernetes.io/projected/5edce1f1-9c11-4049-b286-a16939acb2e7-kube-api-access-jvqh6\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.783340 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-config-data\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.784213 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852927ce-1113-484c-a27f-5e9f75c7f7f0-logs\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.784654 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-scripts\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.815434 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/852927ce-1113-484c-a27f-5e9f75c7f7f0-horizon-secret-key\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.820149 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.858494 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thq65\" (UniqueName: \"kubernetes.io/projected/852927ce-1113-484c-a27f-5e9f75c7f7f0-kube-api-access-thq65\") pod \"horizon-b7fdb6f4c-7tm5b\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.892813 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.892919 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvqh6\" (UniqueName: \"kubernetes.io/projected/5edce1f1-9c11-4049-b286-a16939acb2e7-kube-api-access-jvqh6\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.892979 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5edce1f1-9c11-4049-b286-a16939acb2e7-logs\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.893005 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-config-data\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.897080 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-config-data\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.897308 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5edce1f1-9c11-4049-b286-a16939acb2e7-logs\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.926862 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.944480 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvqh6\" (UniqueName: \"kubernetes.io/projected/5edce1f1-9c11-4049-b286-a16939acb2e7-kube-api-access-jvqh6\") pod \"watcher-applier-0\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " pod="openstack/watcher-applier-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.944648 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-72sks"] Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.945751 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.951564 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.956307 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.956771 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 14:09:56 crc kubenswrapper[4677]: I1203 14:09:56.965315 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9nljg" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.001375 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-njl82"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.002801 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.004119 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-combined-ca-bundle\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.004274 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25vw9\" (UniqueName: \"kubernetes.io/projected/d88b54f3-fe9e-4247-8292-1f513f84fd33-kube-api-access-25vw9\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.004388 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-config\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.021026 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.032461 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.032646 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kpbvh" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.032740 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.038552 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.042827 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.054021 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-72sks"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.071402 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.093837 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-njl82"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.103610 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.108786 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.108833 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-db-sync-config-data\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.108880 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-combined-ca-bundle\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.108900 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-config\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.108963 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.108987 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tctg\" (UniqueName: \"kubernetes.io/projected/0fd09b0c-1690-42e1-b97a-a18e7fefc813-kube-api-access-6tctg\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109008 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-scripts\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109049 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fd09b0c-1690-42e1-b97a-a18e7fefc813-etc-machine-id\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109064 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-config-data\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109086 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-combined-ca-bundle\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109121 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-config-data\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109168 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25vw9\" (UniqueName: \"kubernetes.io/projected/d88b54f3-fe9e-4247-8292-1f513f84fd33-kube-api-access-25vw9\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109198 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vffd9\" (UniqueName: \"kubernetes.io/projected/b40fca00-06a6-4109-bfc5-d67649e1a51c-kube-api-access-vffd9\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.109213 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40fca00-06a6-4109-bfc5-d67649e1a51c-logs\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.147792 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-config\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.149562 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25vw9\" (UniqueName: \"kubernetes.io/projected/d88b54f3-fe9e-4247-8292-1f513f84fd33-kube-api-access-25vw9\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.160210 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-combined-ca-bundle\") pod \"neutron-db-sync-72sks\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.206306 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.210864 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.210932 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tctg\" (UniqueName: \"kubernetes.io/projected/0fd09b0c-1690-42e1-b97a-a18e7fefc813-kube-api-access-6tctg\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.210973 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-scripts\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211003 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fd09b0c-1690-42e1-b97a-a18e7fefc813-etc-machine-id\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211026 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-config-data\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211061 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-config-data\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211106 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vffd9\" (UniqueName: \"kubernetes.io/projected/b40fca00-06a6-4109-bfc5-d67649e1a51c-kube-api-access-vffd9\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211123 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40fca00-06a6-4109-bfc5-d67649e1a51c-logs\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211164 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211193 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-db-sync-config-data\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.211234 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-combined-ca-bundle\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.214778 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-config-data\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.220740 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-72sks" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.220846 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fd09b0c-1690-42e1-b97a-a18e7fefc813-etc-machine-id\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.232251 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40fca00-06a6-4109-bfc5-d67649e1a51c-logs\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.235782 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.236301 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-combined-ca-bundle\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.237833 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-db-sync-config-data\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.237846 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-config-data\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.247266 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-scripts\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.278511 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.293362 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tctg\" (UniqueName: \"kubernetes.io/projected/0fd09b0c-1690-42e1-b97a-a18e7fefc813-kube-api-access-6tctg\") pod \"cinder-db-sync-njl82\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.294026 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-njl82" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.300321 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vffd9\" (UniqueName: \"kubernetes.io/projected/b40fca00-06a6-4109-bfc5-d67649e1a51c-kube-api-access-vffd9\") pod \"watcher-decision-engine-0\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.308807 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.309211 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68758c88fc-fgg5d"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.310692 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.324344 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64cbd68bf9-nhcgh"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.378381 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68758c88fc-fgg5d"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.405050 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dcc496dd5-87k8s"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.406893 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.416292 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-hgx4j"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.417553 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djss7\" (UniqueName: \"kubernetes.io/projected/17b02f27-1f55-4ecc-b7c2-f860720857d2-kube-api-access-djss7\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.417617 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-scripts\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.417669 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-config-data\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.417804 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b02f27-1f55-4ecc-b7c2-f860720857d2-logs\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.417825 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17b02f27-1f55-4ecc-b7c2-f860720857d2-horizon-secret-key\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.420383 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.430970 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.433073 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hgx4j"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.434608 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d2zwm" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.446735 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcc496dd5-87k8s"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.465178 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-f74nz"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.467635 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.470582 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qrxb8" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.470764 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.471210 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.476275 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-f74nz"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.529315 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.532374 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.535457 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.535570 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.535799 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djss7\" (UniqueName: \"kubernetes.io/projected/17b02f27-1f55-4ecc-b7c2-f860720857d2-kube-api-access-djss7\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.536276 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-config\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.536413 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-scripts\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.536520 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb7l9\" (UniqueName: \"kubernetes.io/projected/25625b10-c45e-40f0-b342-bdcdc3e79c9c-kube-api-access-pb7l9\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.536634 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-swift-storage-0\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.536737 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-svc\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.536849 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-sb\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537016 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-scripts\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537138 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-config-data\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537316 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-combined-ca-bundle\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537434 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-nb\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537543 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-logs\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537683 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxdr4\" (UniqueName: \"kubernetes.io/projected/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-kube-api-access-zxdr4\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537811 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-db-sync-config-data\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537940 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b02f27-1f55-4ecc-b7c2-f860720857d2-logs\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.538047 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17b02f27-1f55-4ecc-b7c2-f860720857d2-horizon-secret-key\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.538132 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-config-data\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.538200 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-combined-ca-bundle\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.538303 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77nq8\" (UniqueName: \"kubernetes.io/projected/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-kube-api-access-77nq8\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.537347 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-scripts\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.545050 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b02f27-1f55-4ecc-b7c2-f860720857d2-logs\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.546883 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17b02f27-1f55-4ecc-b7c2-f860720857d2-horizon-secret-key\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.550275 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.552167 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-config-data\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.562836 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djss7\" (UniqueName: \"kubernetes.io/projected/17b02f27-1f55-4ecc-b7c2-f860720857d2-kube-api-access-djss7\") pod \"horizon-68758c88fc-fgg5d\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640654 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-combined-ca-bundle\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640690 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-config-data\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640725 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77nq8\" (UniqueName: \"kubernetes.io/projected/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-kube-api-access-77nq8\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640774 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-config\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640796 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb7l9\" (UniqueName: \"kubernetes.io/projected/25625b10-c45e-40f0-b342-bdcdc3e79c9c-kube-api-access-pb7l9\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640814 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-swift-storage-0\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640838 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-svc\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640862 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-sb\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640891 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-scripts\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640936 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-nb\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640965 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-combined-ca-bundle\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.640987 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-logs\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.641033 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxdr4\" (UniqueName: \"kubernetes.io/projected/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-kube-api-access-zxdr4\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.641067 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-db-sync-config-data\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.641996 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-swift-storage-0\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.643160 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-svc\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.643813 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-sb\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.643876 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-nb\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.647977 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-logs\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.648911 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-config\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.649835 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-db-sync-config-data\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.650798 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-config-data\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.658757 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-combined-ca-bundle\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.668072 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-scripts\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.670472 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-combined-ca-bundle\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.671001 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.678542 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77nq8\" (UniqueName: \"kubernetes.io/projected/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-kube-api-access-77nq8\") pod \"dnsmasq-dns-dcc496dd5-87k8s\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.679007 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxdr4\" (UniqueName: \"kubernetes.io/projected/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-kube-api-access-zxdr4\") pod \"placement-db-sync-f74nz\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.700904 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb7l9\" (UniqueName: \"kubernetes.io/projected/25625b10-c45e-40f0-b342-bdcdc3e79c9c-kube-api-access-pb7l9\") pod \"barbican-db-sync-hgx4j\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.728880 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8r7lf"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.747882 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hglx2\" (UniqueName: \"kubernetes.io/projected/2f345427-301d-44b0-8e50-3d5e2444a48f-kube-api-access-hglx2\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.748131 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-run-httpd\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.748179 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-config-data\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.748203 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.748290 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-log-httpd\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.748346 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.748377 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-scripts\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.764377 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.825693 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.855385 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hglx2\" (UniqueName: \"kubernetes.io/projected/2f345427-301d-44b0-8e50-3d5e2444a48f-kube-api-access-hglx2\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.855475 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-run-httpd\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.855498 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-config-data\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.855513 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.855548 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-log-httpd\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.855574 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.855591 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-scripts\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.859043 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64cbd68bf9-nhcgh"] Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.862766 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-log-httpd\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.882899 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f74nz" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.889590 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.893991 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-scripts\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.894005 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-config-data\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.896098 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-run-httpd\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.898839 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hglx2\" (UniqueName: \"kubernetes.io/projected/2f345427-301d-44b0-8e50-3d5e2444a48f-kube-api-access-hglx2\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.899611 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " pod="openstack/ceilometer-0" Dec 03 14:09:57 crc kubenswrapper[4677]: I1203 14:09:57.924524 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.042116 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b7fdb6f4c-7tm5b"] Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.199898 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.265419 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-njl82"] Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.297105 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.305093 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.319283 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-72sks"] Dec 03 14:09:58 crc kubenswrapper[4677]: W1203 14:09:58.325587 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40fca00_06a6_4109_bfc5_d67649e1a51c.slice/crio-7bc24f74215cf17a781c27d148298785755852661db2aeee03fcab1f522273dd WatchSource:0}: Error finding container 7bc24f74215cf17a781c27d148298785755852661db2aeee03fcab1f522273dd: Status 404 returned error can't find the container with id 7bc24f74215cf17a781c27d148298785755852661db2aeee03fcab1f522273dd Dec 03 14:09:58 crc kubenswrapper[4677]: W1203 14:09:58.345303 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5edce1f1_9c11_4049_b286_a16939acb2e7.slice/crio-23b01ff80ae7cb01f2e27c8da418d73a2ade4176ea17e83f6742857024b82166 WatchSource:0}: Error finding container 23b01ff80ae7cb01f2e27c8da418d73a2ade4176ea17e83f6742857024b82166: Status 404 returned error can't find the container with id 23b01ff80ae7cb01f2e27c8da418d73a2ade4176ea17e83f6742857024b82166 Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.362465 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"b40fca00-06a6-4109-bfc5-d67649e1a51c","Type":"ContainerStarted","Data":"7bc24f74215cf17a781c27d148298785755852661db2aeee03fcab1f522273dd"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.364404 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"5b533531-636f-4285-9c80-996c036eb4b1","Type":"ContainerStarted","Data":"5fde2be1535087fc3b2cf6540a0587d8939607770b726687c369ba2c02fa880b"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.366602 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b7fdb6f4c-7tm5b" event={"ID":"852927ce-1113-484c-a27f-5e9f75c7f7f0","Type":"ContainerStarted","Data":"e7829c2bedddd0678dbefc1b579b9e86f287ffaf8e525897cd0b200a1200ae04"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.369203 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-72sks" event={"ID":"d88b54f3-fe9e-4247-8292-1f513f84fd33","Type":"ContainerStarted","Data":"b4d5e022a662596ddae59eda08164f8be2af48f132d618e8479e36e0c8646a63"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.372747 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" event={"ID":"199679f6-a558-44e5-bd68-cc19b27a2f85","Type":"ContainerStarted","Data":"182805c96102f46bc6c8f5169c72c9fe17b8206a249985f6f2baf501d98a30ea"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.379588 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8r7lf" event={"ID":"d64c9f49-53ee-4069-af26-e2e01ead2ff3","Type":"ContainerStarted","Data":"1b877a7091dd9c860b5104501ed4c0d071447f1de0bc8d1718d3ab84cfd3e556"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.379643 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8r7lf" event={"ID":"d64c9f49-53ee-4069-af26-e2e01ead2ff3","Type":"ContainerStarted","Data":"be20714c847c2e33f67ff1ee2939a87c690cccfa162445687aa5dbdfb3f24f22"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.384017 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-njl82" event={"ID":"0fd09b0c-1690-42e1-b97a-a18e7fefc813","Type":"ContainerStarted","Data":"f99039f40832f2e19ff5f6dfbade0ff5509753b5d47c31912f7113c43996b123"} Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.421772 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8r7lf" podStartSLOduration=2.421748298 podStartE2EDuration="2.421748298s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:58.402849396 +0000 UTC m=+1389.149181851" watchObservedRunningTime="2025-12-03 14:09:58.421748298 +0000 UTC m=+1389.168080763" Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.565157 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcc496dd5-87k8s"] Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.573234 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68758c88fc-fgg5d"] Dec 03 14:09:58 crc kubenswrapper[4677]: W1203 14:09:58.595149 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17b02f27_1f55_4ecc_b7c2_f860720857d2.slice/crio-41275cc6199fc83c0e792cc009bd5ffc308c1b0d085d6aecd90c0e137137fa40 WatchSource:0}: Error finding container 41275cc6199fc83c0e792cc009bd5ffc308c1b0d085d6aecd90c0e137137fa40: Status 404 returned error can't find the container with id 41275cc6199fc83c0e792cc009bd5ffc308c1b0d085d6aecd90c0e137137fa40 Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.756767 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hgx4j"] Dec 03 14:09:58 crc kubenswrapper[4677]: I1203 14:09:58.826127 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-f74nz"] Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.019820 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.293535 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.377095 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68758c88fc-fgg5d"] Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.442203 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.472233 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-976df9d9-dcx8g"] Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.474276 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.485631 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-976df9d9-dcx8g"] Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.501549 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-72sks" event={"ID":"d88b54f3-fe9e-4247-8292-1f513f84fd33","Type":"ContainerStarted","Data":"bfe141487a2b8cfff701b2ecbb5a27bcd223a4147c90788060d261ba646c6a83"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.530466 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5edce1f1-9c11-4049-b286-a16939acb2e7","Type":"ContainerStarted","Data":"23b01ff80ae7cb01f2e27c8da418d73a2ade4176ea17e83f6742857024b82166"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.539278 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17405477-0e11-40ac-b33d-b1b614827780-horizon-secret-key\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.539371 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-scripts\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.539444 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2rpf\" (UniqueName: \"kubernetes.io/projected/17405477-0e11-40ac-b33d-b1b614827780-kube-api-access-v2rpf\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.539473 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17405477-0e11-40ac-b33d-b1b614827780-logs\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.539491 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-config-data\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.581984 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-72sks" podStartSLOduration=3.581966135 podStartE2EDuration="3.581966135s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:59.555433016 +0000 UTC m=+1390.301765491" watchObservedRunningTime="2025-12-03 14:09:59.581966135 +0000 UTC m=+1390.328298590" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.590228 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"5b533531-636f-4285-9c80-996c036eb4b1","Type":"ContainerStarted","Data":"54bde3200fb30d3600edb18d2265f2ec1bd5e215d84a37a590b65bc099fa43a3"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.590282 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"5b533531-636f-4285-9c80-996c036eb4b1","Type":"ContainerStarted","Data":"f754ecd4d121f270a28d322a0c1dfc8b5a8dc2cfc26968d7adfe3d3f99188ebd"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.591227 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.608283 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hgx4j" event={"ID":"25625b10-c45e-40f0-b342-bdcdc3e79c9c","Type":"ContainerStarted","Data":"0e8086cca4d17a62b570dc85b637b50f072e46f99feb791e5c32cd1c611f2cfc"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.623321 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68758c88fc-fgg5d" event={"ID":"17b02f27-1f55-4ecc-b7c2-f860720857d2","Type":"ContainerStarted","Data":"41275cc6199fc83c0e792cc009bd5ffc308c1b0d085d6aecd90c0e137137fa40"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.626606 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerStarted","Data":"9e7f7fe92ce681f771cb76ea336c3d9d35096f5c7f447611daa512ad35854582"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.628747 4677 generic.go:334] "Generic (PLEG): container finished" podID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerID="900ccc4aa90c89810a4522a43cd33ff556c623eccea11cc243aae76dc4da172f" exitCode=0 Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.628859 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" event={"ID":"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37","Type":"ContainerDied","Data":"900ccc4aa90c89810a4522a43cd33ff556c623eccea11cc243aae76dc4da172f"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.628879 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" event={"ID":"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37","Type":"ContainerStarted","Data":"cd637071b474d2843176fabf5f6e40a3770e98e4a725cbeb4b0cab40cf44e3ed"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.641096 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17405477-0e11-40ac-b33d-b1b614827780-logs\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.641152 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-config-data\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.641220 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17405477-0e11-40ac-b33d-b1b614827780-horizon-secret-key\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.641284 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-scripts\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.641347 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2rpf\" (UniqueName: \"kubernetes.io/projected/17405477-0e11-40ac-b33d-b1b614827780-kube-api-access-v2rpf\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.641783 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17405477-0e11-40ac-b33d-b1b614827780-logs\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.643474 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-scripts\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.647153 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f74nz" event={"ID":"24ec7657-9bc6-4953-9b8c-d07a62b7aff1","Type":"ContainerStarted","Data":"e3f6ff5bbdc3c9f850810a53e188cdd38088026399f295d9025a5b3e90a8fa7a"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.647813 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=3.647794558 podStartE2EDuration="3.647794558s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:09:59.636210302 +0000 UTC m=+1390.382542767" watchObservedRunningTime="2025-12-03 14:09:59.647794558 +0000 UTC m=+1390.394127013" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.650086 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-config-data\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.661232 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17405477-0e11-40ac-b33d-b1b614827780-horizon-secret-key\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.669179 4677 generic.go:334] "Generic (PLEG): container finished" podID="199679f6-a558-44e5-bd68-cc19b27a2f85" containerID="d3101e0a827beec899b0423c9fad39e0b10d692b2a32edf41c296b72fc75a1e1" exitCode=0 Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.670081 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" event={"ID":"199679f6-a558-44e5-bd68-cc19b27a2f85","Type":"ContainerDied","Data":"d3101e0a827beec899b0423c9fad39e0b10d692b2a32edf41c296b72fc75a1e1"} Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.702016 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2rpf\" (UniqueName: \"kubernetes.io/projected/17405477-0e11-40ac-b33d-b1b614827780-kube-api-access-v2rpf\") pod \"horizon-976df9d9-dcx8g\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:09:59 crc kubenswrapper[4677]: I1203 14:09:59.829808 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.305247 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.485987 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-sb\") pod \"199679f6-a558-44e5-bd68-cc19b27a2f85\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.486046 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-config\") pod \"199679f6-a558-44e5-bd68-cc19b27a2f85\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.486090 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-nb\") pod \"199679f6-a558-44e5-bd68-cc19b27a2f85\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.486151 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf54x\" (UniqueName: \"kubernetes.io/projected/199679f6-a558-44e5-bd68-cc19b27a2f85-kube-api-access-wf54x\") pod \"199679f6-a558-44e5-bd68-cc19b27a2f85\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.486224 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-svc\") pod \"199679f6-a558-44e5-bd68-cc19b27a2f85\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.486300 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-swift-storage-0\") pod \"199679f6-a558-44e5-bd68-cc19b27a2f85\" (UID: \"199679f6-a558-44e5-bd68-cc19b27a2f85\") " Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.517588 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "199679f6-a558-44e5-bd68-cc19b27a2f85" (UID: "199679f6-a558-44e5-bd68-cc19b27a2f85"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.517764 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/199679f6-a558-44e5-bd68-cc19b27a2f85-kube-api-access-wf54x" (OuterVolumeSpecName: "kube-api-access-wf54x") pod "199679f6-a558-44e5-bd68-cc19b27a2f85" (UID: "199679f6-a558-44e5-bd68-cc19b27a2f85"). InnerVolumeSpecName "kube-api-access-wf54x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.538166 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "199679f6-a558-44e5-bd68-cc19b27a2f85" (UID: "199679f6-a558-44e5-bd68-cc19b27a2f85"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.564112 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "199679f6-a558-44e5-bd68-cc19b27a2f85" (UID: "199679f6-a558-44e5-bd68-cc19b27a2f85"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.566167 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "199679f6-a558-44e5-bd68-cc19b27a2f85" (UID: "199679f6-a558-44e5-bd68-cc19b27a2f85"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.584243 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-config" (OuterVolumeSpecName: "config") pod "199679f6-a558-44e5-bd68-cc19b27a2f85" (UID: "199679f6-a558-44e5-bd68-cc19b27a2f85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.588592 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.588611 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.588621 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.588630 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.588640 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf54x\" (UniqueName: \"kubernetes.io/projected/199679f6-a558-44e5-bd68-cc19b27a2f85-kube-api-access-wf54x\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.588649 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199679f6-a558-44e5-bd68-cc19b27a2f85-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.616773 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-976df9d9-dcx8g"] Dec 03 14:10:01 crc kubenswrapper[4677]: W1203 14:10:00.651486 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17405477_0e11_40ac_b33d_b1b614827780.slice/crio-ca97e40f9c7d0553e8703a3acd14d40194c3adf5c07ee4405472827ce402fd08 WatchSource:0}: Error finding container ca97e40f9c7d0553e8703a3acd14d40194c3adf5c07ee4405472827ce402fd08: Status 404 returned error can't find the container with id ca97e40f9c7d0553e8703a3acd14d40194c3adf5c07ee4405472827ce402fd08 Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.701402 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-976df9d9-dcx8g" event={"ID":"17405477-0e11-40ac-b33d-b1b614827780","Type":"ContainerStarted","Data":"ca97e40f9c7d0553e8703a3acd14d40194c3adf5c07ee4405472827ce402fd08"} Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.710438 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" event={"ID":"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37","Type":"ContainerStarted","Data":"485c37947f3b1ab828005255c2341513bc22935bc837f4ede0426514259fadd1"} Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.710525 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.721459 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.722527 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cbd68bf9-nhcgh" event={"ID":"199679f6-a558-44e5-bd68-cc19b27a2f85","Type":"ContainerDied","Data":"182805c96102f46bc6c8f5169c72c9fe17b8206a249985f6f2baf501d98a30ea"} Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.722597 4677 scope.go:117] "RemoveContainer" containerID="d3101e0a827beec899b0423c9fad39e0b10d692b2a32edf41c296b72fc75a1e1" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.722778 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api-log" containerID="cri-o://f754ecd4d121f270a28d322a0c1dfc8b5a8dc2cfc26968d7adfe3d3f99188ebd" gracePeriod=30 Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.722818 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" containerID="cri-o://54bde3200fb30d3600edb18d2265f2ec1bd5e215d84a37a590b65bc099fa43a3" gracePeriod=30 Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.737637 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.145:9322/\": EOF" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.741247 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" podStartSLOduration=4.741223825 podStartE2EDuration="4.741223825s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:00.7335723 +0000 UTC m=+1391.479904785" watchObservedRunningTime="2025-12-03 14:10:00.741223825 +0000 UTC m=+1391.487556290" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.744382 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.145:9322/\": EOF" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.822620 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64cbd68bf9-nhcgh"] Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:00.838556 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64cbd68bf9-nhcgh"] Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:01.767768 4677 generic.go:334] "Generic (PLEG): container finished" podID="5b533531-636f-4285-9c80-996c036eb4b1" containerID="f754ecd4d121f270a28d322a0c1dfc8b5a8dc2cfc26968d7adfe3d3f99188ebd" exitCode=143 Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:01.768671 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"5b533531-636f-4285-9c80-996c036eb4b1","Type":"ContainerDied","Data":"f754ecd4d121f270a28d322a0c1dfc8b5a8dc2cfc26968d7adfe3d3f99188ebd"} Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:01.952422 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 14:10:01 crc kubenswrapper[4677]: I1203 14:10:01.987520 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="199679f6-a558-44e5-bd68-cc19b27a2f85" path="/var/lib/kubelet/pods/199679f6-a558-44e5-bd68-cc19b27a2f85/volumes" Dec 03 14:10:04 crc kubenswrapper[4677]: I1203 14:10:04.889686 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.145:9322/\": read tcp 10.217.0.2:44694->10.217.0.145:9322: read: connection reset by peer" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.505694 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b7fdb6f4c-7tm5b"] Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.543456 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85b5c78d96-8262j"] Dec 03 14:10:05 crc kubenswrapper[4677]: E1203 14:10:05.543870 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199679f6-a558-44e5-bd68-cc19b27a2f85" containerName="init" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.543885 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="199679f6-a558-44e5-bd68-cc19b27a2f85" containerName="init" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.544068 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="199679f6-a558-44e5-bd68-cc19b27a2f85" containerName="init" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.545163 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.553356 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.578713 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85b5c78d96-8262j"] Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.593309 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-tls-certs\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.593367 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ec3794-7d5b-49dc-a957-bee4761d55f0-logs\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.593395 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-combined-ca-bundle\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.593419 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-secret-key\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.593439 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2f5n\" (UniqueName: \"kubernetes.io/projected/c4ec3794-7d5b-49dc-a957-bee4761d55f0-kube-api-access-c2f5n\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.593507 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-config-data\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.593532 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-scripts\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.636982 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-976df9d9-dcx8g"] Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.670020 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c9d4fbd46-n2z95"] Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.672113 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.690236 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c9d4fbd46-n2z95"] Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.695992 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-combined-ca-bundle\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696076 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-horizon-tls-certs\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696104 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-tls-certs\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696136 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ec3794-7d5b-49dc-a957-bee4761d55f0-logs\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696163 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-combined-ca-bundle\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696181 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-horizon-secret-key\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696199 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-secret-key\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696217 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2f5n\" (UniqueName: \"kubernetes.io/projected/c4ec3794-7d5b-49dc-a957-bee4761d55f0-kube-api-access-c2f5n\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696248 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njh2f\" (UniqueName: \"kubernetes.io/projected/0749a528-b8b2-497f-bde8-f7b7765e8068-kube-api-access-njh2f\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696267 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0749a528-b8b2-497f-bde8-f7b7765e8068-scripts\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696308 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-config-data\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696323 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0749a528-b8b2-497f-bde8-f7b7765e8068-config-data\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696344 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-scripts\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696364 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0749a528-b8b2-497f-bde8-f7b7765e8068-logs\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.696735 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ec3794-7d5b-49dc-a957-bee4761d55f0-logs\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.705027 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-config-data\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.705438 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-secret-key\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.706206 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-scripts\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.718140 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-combined-ca-bundle\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.718417 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-tls-certs\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.725178 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2f5n\" (UniqueName: \"kubernetes.io/projected/c4ec3794-7d5b-49dc-a957-bee4761d55f0-kube-api-access-c2f5n\") pod \"horizon-85b5c78d96-8262j\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798182 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-horizon-secret-key\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798262 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njh2f\" (UniqueName: \"kubernetes.io/projected/0749a528-b8b2-497f-bde8-f7b7765e8068-kube-api-access-njh2f\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798284 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0749a528-b8b2-497f-bde8-f7b7765e8068-scripts\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798333 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0749a528-b8b2-497f-bde8-f7b7765e8068-config-data\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798360 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0749a528-b8b2-497f-bde8-f7b7765e8068-logs\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798396 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-combined-ca-bundle\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798464 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-horizon-tls-certs\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.798912 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0749a528-b8b2-497f-bde8-f7b7765e8068-logs\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.799270 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0749a528-b8b2-497f-bde8-f7b7765e8068-scripts\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.800076 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0749a528-b8b2-497f-bde8-f7b7765e8068-config-data\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.801445 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-horizon-secret-key\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.803619 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-horizon-tls-certs\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.804993 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0749a528-b8b2-497f-bde8-f7b7765e8068-combined-ca-bundle\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.817180 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njh2f\" (UniqueName: \"kubernetes.io/projected/0749a528-b8b2-497f-bde8-f7b7765e8068-kube-api-access-njh2f\") pod \"horizon-c9d4fbd46-n2z95\" (UID: \"0749a528-b8b2-497f-bde8-f7b7765e8068\") " pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.830902 4677 generic.go:334] "Generic (PLEG): container finished" podID="5b533531-636f-4285-9c80-996c036eb4b1" containerID="54bde3200fb30d3600edb18d2265f2ec1bd5e215d84a37a590b65bc099fa43a3" exitCode=0 Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.830942 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"5b533531-636f-4285-9c80-996c036eb4b1","Type":"ContainerDied","Data":"54bde3200fb30d3600edb18d2265f2ec1bd5e215d84a37a590b65bc099fa43a3"} Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.869864 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:05 crc kubenswrapper[4677]: I1203 14:10:05.993593 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:06 crc kubenswrapper[4677]: I1203 14:10:06.952783 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.145:9322/\": dial tcp 10.217.0.145:9322: connect: connection refused" Dec 03 14:10:07 crc kubenswrapper[4677]: I1203 14:10:07.765698 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:10:07 crc kubenswrapper[4677]: I1203 14:10:07.836705 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj"] Dec 03 14:10:07 crc kubenswrapper[4677]: I1203 14:10:07.837035 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" containerID="cri-o://ae5ab29a0972e40bec375a1634edb74351d79e1fd5e1d4ec70437460e2232b5c" gracePeriod=10 Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.437774 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.437839 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.437889 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.438703 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"960b2bbd5800fa0b1e3c63b4bd748d2c40afb2ceea7aa3529320dbcc66c17398"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.438958 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://960b2bbd5800fa0b1e3c63b4bd748d2c40afb2ceea7aa3529320dbcc66c17398" gracePeriod=600 Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.878937 4677 generic.go:334] "Generic (PLEG): container finished" podID="3860befc-70ca-4658-b967-b3e8351c9788" containerID="ae5ab29a0972e40bec375a1634edb74351d79e1fd5e1d4ec70437460e2232b5c" exitCode=0 Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.878989 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" event={"ID":"3860befc-70ca-4658-b967-b3e8351c9788","Type":"ContainerDied","Data":"ae5ab29a0972e40bec375a1634edb74351d79e1fd5e1d4ec70437460e2232b5c"} Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.883044 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="960b2bbd5800fa0b1e3c63b4bd748d2c40afb2ceea7aa3529320dbcc66c17398" exitCode=0 Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.883079 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"960b2bbd5800fa0b1e3c63b4bd748d2c40afb2ceea7aa3529320dbcc66c17398"} Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.883133 4677 scope.go:117] "RemoveContainer" containerID="2e1aec0429580d81a4c5a21902c4598063a10937c54a8f7496939174766a6f58" Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.886285 4677 generic.go:334] "Generic (PLEG): container finished" podID="d64c9f49-53ee-4069-af26-e2e01ead2ff3" containerID="1b877a7091dd9c860b5104501ed4c0d071447f1de0bc8d1718d3ab84cfd3e556" exitCode=0 Dec 03 14:10:08 crc kubenswrapper[4677]: I1203 14:10:08.886328 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8r7lf" event={"ID":"d64c9f49-53ee-4069-af26-e2e01ead2ff3","Type":"ContainerDied","Data":"1b877a7091dd9c860b5104501ed4c0d071447f1de0bc8d1718d3ab84cfd3e556"} Dec 03 14:10:09 crc kubenswrapper[4677]: I1203 14:10:09.290718 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Dec 03 14:10:13 crc kubenswrapper[4677]: I1203 14:10:13.949297 4677 generic.go:334] "Generic (PLEG): container finished" podID="4c12c954-d94a-404a-9263-a8371fdc12e4" containerID="8c5473373d1e776782a3579f5497a7aee84971d76b30289c6cc605d85f68a020" exitCode=0 Dec 03 14:10:13 crc kubenswrapper[4677]: I1203 14:10:13.949846 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbkxk" event={"ID":"4c12c954-d94a-404a-9263-a8371fdc12e4","Type":"ContainerDied","Data":"8c5473373d1e776782a3579f5497a7aee84971d76b30289c6cc605d85f68a020"} Dec 03 14:10:16 crc kubenswrapper[4677]: I1203 14:10:16.953769 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.145:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.595440 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.595753 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.595866 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc6h98h559h5b7h657h95hf4h65bh55fh4hfh67h596h5bbh574h55bh55h689h695h686h88h675h65dh677hc7h696h5d9hb6h55ch64h66ch8cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thq65,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-b7fdb6f4c-7tm5b_openstack(852927ce-1113-484c-a27f-5e9f75c7f7f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.603709 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-b7fdb6f4c-7tm5b" podUID="852927ce-1113-484c-a27f-5e9f75c7f7f0" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.605596 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.605630 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.605739 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n699h64fh687hch5b6h566h54ch7bh55fh5dch5f7h58dhd6h64dh559h557hb7h65fh686h65bh599h567h645h6fhb4h76hfdh5ffh64ch564hbbh64bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v2rpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-976df9d9-dcx8g_openstack(17405477-0e11-40ac-b33d-b1b614827780): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.617797 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-976df9d9-dcx8g" podUID="17405477-0e11-40ac-b33d-b1b614827780" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.649437 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.649487 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.649660 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdhd8h58bh67fh579h5dbh685h87h8dhb5h95h6fh557h59bhd6h84h575h546h8h9fh85h56fhf7h64dh66fh585h5dbh8dh549h548h668h5fbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-djss7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-68758c88fc-fgg5d_openstack(17b02f27-1f55-4ecc-b7c2-f860720857d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:10:18 crc kubenswrapper[4677]: E1203 14:10:18.651858 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-68758c88fc-fgg5d" podUID="17b02f27-1f55-4ecc-b7c2-f860720857d2" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.730719 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.754523 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-config-data\") pod \"5b533531-636f-4285-9c80-996c036eb4b1\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.754812 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b533531-636f-4285-9c80-996c036eb4b1-logs\") pod \"5b533531-636f-4285-9c80-996c036eb4b1\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.754982 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-custom-prometheus-ca\") pod \"5b533531-636f-4285-9c80-996c036eb4b1\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.755126 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-combined-ca-bundle\") pod \"5b533531-636f-4285-9c80-996c036eb4b1\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.755358 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf8zn\" (UniqueName: \"kubernetes.io/projected/5b533531-636f-4285-9c80-996c036eb4b1-kube-api-access-jf8zn\") pod \"5b533531-636f-4285-9c80-996c036eb4b1\" (UID: \"5b533531-636f-4285-9c80-996c036eb4b1\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.757325 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b533531-636f-4285-9c80-996c036eb4b1-logs" (OuterVolumeSpecName: "logs") pod "5b533531-636f-4285-9c80-996c036eb4b1" (UID: "5b533531-636f-4285-9c80-996c036eb4b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.759081 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.760126 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b533531-636f-4285-9c80-996c036eb4b1-kube-api-access-jf8zn" (OuterVolumeSpecName: "kube-api-access-jf8zn") pod "5b533531-636f-4285-9c80-996c036eb4b1" (UID: "5b533531-636f-4285-9c80-996c036eb4b1"). InnerVolumeSpecName "kube-api-access-jf8zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.794390 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b533531-636f-4285-9c80-996c036eb4b1" (UID: "5b533531-636f-4285-9c80-996c036eb4b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.797374 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "5b533531-636f-4285-9c80-996c036eb4b1" (UID: "5b533531-636f-4285-9c80-996c036eb4b1"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.820244 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-config-data" (OuterVolumeSpecName: "config-data") pod "5b533531-636f-4285-9c80-996c036eb4b1" (UID: "5b533531-636f-4285-9c80-996c036eb4b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.856997 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-credential-keys\") pod \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.857050 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-fernet-keys\") pod \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.857079 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65h8c\" (UniqueName: \"kubernetes.io/projected/d64c9f49-53ee-4069-af26-e2e01ead2ff3-kube-api-access-65h8c\") pod \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.857325 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-config-data\") pod \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.857383 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-scripts\") pod \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.857429 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-combined-ca-bundle\") pod \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\" (UID: \"d64c9f49-53ee-4069-af26-e2e01ead2ff3\") " Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.858036 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf8zn\" (UniqueName: \"kubernetes.io/projected/5b533531-636f-4285-9c80-996c036eb4b1-kube-api-access-jf8zn\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.858059 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.858071 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b533531-636f-4285-9c80-996c036eb4b1-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.858083 4677 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.858093 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b533531-636f-4285-9c80-996c036eb4b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.860391 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d64c9f49-53ee-4069-af26-e2e01ead2ff3" (UID: "d64c9f49-53ee-4069-af26-e2e01ead2ff3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.860469 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d64c9f49-53ee-4069-af26-e2e01ead2ff3" (UID: "d64c9f49-53ee-4069-af26-e2e01ead2ff3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.862808 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64c9f49-53ee-4069-af26-e2e01ead2ff3-kube-api-access-65h8c" (OuterVolumeSpecName: "kube-api-access-65h8c") pod "d64c9f49-53ee-4069-af26-e2e01ead2ff3" (UID: "d64c9f49-53ee-4069-af26-e2e01ead2ff3"). InnerVolumeSpecName "kube-api-access-65h8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.863739 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-scripts" (OuterVolumeSpecName: "scripts") pod "d64c9f49-53ee-4069-af26-e2e01ead2ff3" (UID: "d64c9f49-53ee-4069-af26-e2e01ead2ff3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.886771 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d64c9f49-53ee-4069-af26-e2e01ead2ff3" (UID: "d64c9f49-53ee-4069-af26-e2e01ead2ff3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.890598 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-config-data" (OuterVolumeSpecName: "config-data") pod "d64c9f49-53ee-4069-af26-e2e01ead2ff3" (UID: "d64c9f49-53ee-4069-af26-e2e01ead2ff3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.959736 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.959772 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.959783 4677 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.959791 4677 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.959799 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65h8c\" (UniqueName: \"kubernetes.io/projected/d64c9f49-53ee-4069-af26-e2e01ead2ff3-kube-api-access-65h8c\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:18 crc kubenswrapper[4677]: I1203 14:10:18.959808 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64c9f49-53ee-4069-af26-e2e01ead2ff3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.023433 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8r7lf" event={"ID":"d64c9f49-53ee-4069-af26-e2e01ead2ff3","Type":"ContainerDied","Data":"be20714c847c2e33f67ff1ee2939a87c690cccfa162445687aa5dbdfb3f24f22"} Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.023484 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be20714c847c2e33f67ff1ee2939a87c690cccfa162445687aa5dbdfb3f24f22" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.023535 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8r7lf" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.026287 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.028614 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"5b533531-636f-4285-9c80-996c036eb4b1","Type":"ContainerDied","Data":"5fde2be1535087fc3b2cf6540a0587d8939607770b726687c369ba2c02fa880b"} Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.126398 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.159157 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.185962 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:19 crc kubenswrapper[4677]: E1203 14:10:19.186378 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.186397 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" Dec 03 14:10:19 crc kubenswrapper[4677]: E1203 14:10:19.186415 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64c9f49-53ee-4069-af26-e2e01ead2ff3" containerName="keystone-bootstrap" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.186421 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64c9f49-53ee-4069-af26-e2e01ead2ff3" containerName="keystone-bootstrap" Dec 03 14:10:19 crc kubenswrapper[4677]: E1203 14:10:19.186434 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api-log" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.186440 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api-log" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.186599 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.186621 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64c9f49-53ee-4069-af26-e2e01ead2ff3" containerName="keystone-bootstrap" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.186635 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api-log" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.187570 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.189992 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.195088 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.267541 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e9739b2-3472-4adb-9917-20b334a308cb-logs\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.267935 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k97j\" (UniqueName: \"kubernetes.io/projected/1e9739b2-3472-4adb-9917-20b334a308cb-kube-api-access-9k97j\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.267992 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-config-data\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.268099 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.268184 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.290333 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.369216 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k97j\" (UniqueName: \"kubernetes.io/projected/1e9739b2-3472-4adb-9917-20b334a308cb-kube-api-access-9k97j\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.369269 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-config-data\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.369311 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.369342 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.369381 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e9739b2-3472-4adb-9917-20b334a308cb-logs\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.369868 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e9739b2-3472-4adb-9917-20b334a308cb-logs\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.374497 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.379144 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-config-data\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.381248 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.384943 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k97j\" (UniqueName: \"kubernetes.io/projected/1e9739b2-3472-4adb-9917-20b334a308cb-kube-api-access-9k97j\") pod \"watcher-api-0\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.506405 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.941662 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8r7lf"] Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.950563 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8r7lf"] Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.986518 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b533531-636f-4285-9c80-996c036eb4b1" path="/var/lib/kubelet/pods/5b533531-636f-4285-9c80-996c036eb4b1/volumes" Dec 03 14:10:19 crc kubenswrapper[4677]: I1203 14:10:19.987419 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64c9f49-53ee-4069-af26-e2e01ead2ff3" path="/var/lib/kubelet/pods/d64c9f49-53ee-4069-af26-e2e01ead2ff3/volumes" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.043622 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ff56g"] Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.045915 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.048318 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.048749 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.048890 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.049059 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s8xkt" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.051771 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.068622 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ff56g"] Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.076759 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-scripts\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.076867 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbpw5\" (UniqueName: \"kubernetes.io/projected/47d54aae-4e2b-454f-a29d-9b252fd60179-kube-api-access-mbpw5\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.076898 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-combined-ca-bundle\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.076990 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-fernet-keys\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.077040 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-config-data\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.077128 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-credential-keys\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.178649 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbpw5\" (UniqueName: \"kubernetes.io/projected/47d54aae-4e2b-454f-a29d-9b252fd60179-kube-api-access-mbpw5\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.178715 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-combined-ca-bundle\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.178782 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-fernet-keys\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.178843 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-config-data\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.178884 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-credential-keys\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.179055 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-scripts\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.183897 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-config-data\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.184339 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-combined-ca-bundle\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.184933 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-scripts\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.185156 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-credential-keys\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.185612 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-fernet-keys\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.198540 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbpw5\" (UniqueName: \"kubernetes.io/projected/47d54aae-4e2b-454f-a29d-9b252fd60179-kube-api-access-mbpw5\") pod \"keystone-bootstrap-ff56g\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:20 crc kubenswrapper[4677]: I1203 14:10:20.378588 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:21 crc kubenswrapper[4677]: I1203 14:10:21.954892 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="5b533531-636f-4285-9c80-996c036eb4b1" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.145:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:10:24 crc kubenswrapper[4677]: I1203 14:10:24.291428 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 03 14:10:24 crc kubenswrapper[4677]: I1203 14:10:24.292646 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:10:29 crc kubenswrapper[4677]: I1203 14:10:29.292914 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 03 14:10:30 crc kubenswrapper[4677]: I1203 14:10:30.147304 4677 generic.go:334] "Generic (PLEG): container finished" podID="d88b54f3-fe9e-4247-8292-1f513f84fd33" containerID="bfe141487a2b8cfff701b2ecbb5a27bcd223a4147c90788060d261ba646c6a83" exitCode=0 Dec 03 14:10:30 crc kubenswrapper[4677]: I1203 14:10:30.147389 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-72sks" event={"ID":"d88b54f3-fe9e-4247-8292-1f513f84fd33","Type":"ContainerDied","Data":"bfe141487a2b8cfff701b2ecbb5a27bcd223a4147c90788060d261ba646c6a83"} Dec 03 14:10:31 crc kubenswrapper[4677]: E1203 14:10:31.125682 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Dec 03 14:10:31 crc kubenswrapper[4677]: E1203 14:10:31.125735 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Dec 03 14:10:31 crc kubenswrapper[4677]: E1203 14:10:31.125844 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.129.56.48:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb5h8h9dh649h6bh684h595h576h544h5c8h568hb8h656h5cch5c5h67fh546h5ch5f7hb4h587h647h577hf4h7fh597h657h59fh9ch676h64hc8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hglx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(2f345427-301d-44b0-8e50-3d5e2444a48f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.128311 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:10:31 crc kubenswrapper[4677]: E1203 14:10:31.721570 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Dec 03 14:10:31 crc kubenswrapper[4677]: E1203 14:10:31.721930 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Dec 03 14:10:31 crc kubenswrapper[4677]: E1203 14:10:31.722083 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.129.56.48:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pb7l9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-hgx4j_openstack(25625b10-c45e-40f0-b342-bdcdc3e79c9c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:10:31 crc kubenswrapper[4677]: E1203 14:10:31.723357 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-hgx4j" podUID="25625b10-c45e-40f0-b342-bdcdc3e79c9c" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.847884 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbkxk" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.866339 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.919676 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.924911 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-sb\") pod \"3860befc-70ca-4658-b967-b3e8351c9788\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.924973 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-nb\") pod \"3860befc-70ca-4658-b967-b3e8351c9788\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925004 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-config\") pod \"3860befc-70ca-4658-b967-b3e8351c9788\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925133 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-config-data\") pod \"4c12c954-d94a-404a-9263-a8371fdc12e4\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925189 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-swift-storage-0\") pod \"3860befc-70ca-4658-b967-b3e8351c9788\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925226 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-combined-ca-bundle\") pod \"4c12c954-d94a-404a-9263-a8371fdc12e4\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925329 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt5xq\" (UniqueName: \"kubernetes.io/projected/3860befc-70ca-4658-b967-b3e8351c9788-kube-api-access-kt5xq\") pod \"3860befc-70ca-4658-b967-b3e8351c9788\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925380 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-db-sync-config-data\") pod \"4c12c954-d94a-404a-9263-a8371fdc12e4\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925421 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-svc\") pod \"3860befc-70ca-4658-b967-b3e8351c9788\" (UID: \"3860befc-70ca-4658-b967-b3e8351c9788\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.925454 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq6p8\" (UniqueName: \"kubernetes.io/projected/4c12c954-d94a-404a-9263-a8371fdc12e4-kube-api-access-bq6p8\") pod \"4c12c954-d94a-404a-9263-a8371fdc12e4\" (UID: \"4c12c954-d94a-404a-9263-a8371fdc12e4\") " Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.952349 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4c12c954-d94a-404a-9263-a8371fdc12e4" (UID: "4c12c954-d94a-404a-9263-a8371fdc12e4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.964600 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3860befc-70ca-4658-b967-b3e8351c9788-kube-api-access-kt5xq" (OuterVolumeSpecName: "kube-api-access-kt5xq") pod "3860befc-70ca-4658-b967-b3e8351c9788" (UID: "3860befc-70ca-4658-b967-b3e8351c9788"). InnerVolumeSpecName "kube-api-access-kt5xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.972189 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:10:31 crc kubenswrapper[4677]: I1203 14:10:31.999579 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c12c954-d94a-404a-9263-a8371fdc12e4-kube-api-access-bq6p8" (OuterVolumeSpecName: "kube-api-access-bq6p8") pod "4c12c954-d94a-404a-9263-a8371fdc12e4" (UID: "4c12c954-d94a-404a-9263-a8371fdc12e4"). InnerVolumeSpecName "kube-api-access-bq6p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.032611 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-config-data\") pod \"852927ce-1113-484c-a27f-5e9f75c7f7f0\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.032733 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thq65\" (UniqueName: \"kubernetes.io/projected/852927ce-1113-484c-a27f-5e9f75c7f7f0-kube-api-access-thq65\") pod \"852927ce-1113-484c-a27f-5e9f75c7f7f0\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.032832 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17405477-0e11-40ac-b33d-b1b614827780-horizon-secret-key\") pod \"17405477-0e11-40ac-b33d-b1b614827780\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.032871 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852927ce-1113-484c-a27f-5e9f75c7f7f0-logs\") pod \"852927ce-1113-484c-a27f-5e9f75c7f7f0\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.032890 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17405477-0e11-40ac-b33d-b1b614827780-logs\") pod \"17405477-0e11-40ac-b33d-b1b614827780\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.032937 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-scripts\") pod \"17405477-0e11-40ac-b33d-b1b614827780\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033005 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/852927ce-1113-484c-a27f-5e9f75c7f7f0-horizon-secret-key\") pod \"852927ce-1113-484c-a27f-5e9f75c7f7f0\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033039 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2rpf\" (UniqueName: \"kubernetes.io/projected/17405477-0e11-40ac-b33d-b1b614827780-kube-api-access-v2rpf\") pod \"17405477-0e11-40ac-b33d-b1b614827780\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033107 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-scripts\") pod \"852927ce-1113-484c-a27f-5e9f75c7f7f0\" (UID: \"852927ce-1113-484c-a27f-5e9f75c7f7f0\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033190 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-config-data\") pod \"17405477-0e11-40ac-b33d-b1b614827780\" (UID: \"17405477-0e11-40ac-b33d-b1b614827780\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033367 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852927ce-1113-484c-a27f-5e9f75c7f7f0-logs" (OuterVolumeSpecName: "logs") pod "852927ce-1113-484c-a27f-5e9f75c7f7f0" (UID: "852927ce-1113-484c-a27f-5e9f75c7f7f0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033606 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852927ce-1113-484c-a27f-5e9f75c7f7f0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033628 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt5xq\" (UniqueName: \"kubernetes.io/projected/3860befc-70ca-4658-b967-b3e8351c9788-kube-api-access-kt5xq\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033641 4677 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033652 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq6p8\" (UniqueName: \"kubernetes.io/projected/4c12c954-d94a-404a-9263-a8371fdc12e4-kube-api-access-bq6p8\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.033846 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-config-data" (OuterVolumeSpecName: "config-data") pod "852927ce-1113-484c-a27f-5e9f75c7f7f0" (UID: "852927ce-1113-484c-a27f-5e9f75c7f7f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.041167 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17405477-0e11-40ac-b33d-b1b614827780-logs" (OuterVolumeSpecName: "logs") pod "17405477-0e11-40ac-b33d-b1b614827780" (UID: "17405477-0e11-40ac-b33d-b1b614827780"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.041494 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-scripts" (OuterVolumeSpecName: "scripts") pod "17405477-0e11-40ac-b33d-b1b614827780" (UID: "17405477-0e11-40ac-b33d-b1b614827780"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.047725 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-config-data" (OuterVolumeSpecName: "config-data") pod "17405477-0e11-40ac-b33d-b1b614827780" (UID: "17405477-0e11-40ac-b33d-b1b614827780"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.048645 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-scripts" (OuterVolumeSpecName: "scripts") pod "852927ce-1113-484c-a27f-5e9f75c7f7f0" (UID: "852927ce-1113-484c-a27f-5e9f75c7f7f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.063311 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17405477-0e11-40ac-b33d-b1b614827780-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "17405477-0e11-40ac-b33d-b1b614827780" (UID: "17405477-0e11-40ac-b33d-b1b614827780"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.063478 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852927ce-1113-484c-a27f-5e9f75c7f7f0-kube-api-access-thq65" (OuterVolumeSpecName: "kube-api-access-thq65") pod "852927ce-1113-484c-a27f-5e9f75c7f7f0" (UID: "852927ce-1113-484c-a27f-5e9f75c7f7f0"). InnerVolumeSpecName "kube-api-access-thq65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.079117 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17405477-0e11-40ac-b33d-b1b614827780-kube-api-access-v2rpf" (OuterVolumeSpecName: "kube-api-access-v2rpf") pod "17405477-0e11-40ac-b33d-b1b614827780" (UID: "17405477-0e11-40ac-b33d-b1b614827780"). InnerVolumeSpecName "kube-api-access-v2rpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.101377 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852927ce-1113-484c-a27f-5e9f75c7f7f0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "852927ce-1113-484c-a27f-5e9f75c7f7f0" (UID: "852927ce-1113-484c-a27f-5e9f75c7f7f0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.113139 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c12c954-d94a-404a-9263-a8371fdc12e4" (UID: "4c12c954-d94a-404a-9263-a8371fdc12e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136336 4677 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17405477-0e11-40ac-b33d-b1b614827780-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136379 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17405477-0e11-40ac-b33d-b1b614827780-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136392 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136403 4677 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/852927ce-1113-484c-a27f-5e9f75c7f7f0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136414 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2rpf\" (UniqueName: \"kubernetes.io/projected/17405477-0e11-40ac-b33d-b1b614827780-kube-api-access-v2rpf\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136426 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136437 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17405477-0e11-40ac-b33d-b1b614827780-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136446 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852927ce-1113-484c-a27f-5e9f75c7f7f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136456 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thq65\" (UniqueName: \"kubernetes.io/projected/852927ce-1113-484c-a27f-5e9f75c7f7f0-kube-api-access-thq65\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.136466 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.173212 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.182154 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3860befc-70ca-4658-b967-b3e8351c9788" (UID: "3860befc-70ca-4658-b967-b3e8351c9788"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.186295 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b7fdb6f4c-7tm5b" event={"ID":"852927ce-1113-484c-a27f-5e9f75c7f7f0","Type":"ContainerDied","Data":"e7829c2bedddd0678dbefc1b579b9e86f287ffaf8e525897cd0b200a1200ae04"} Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.186384 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b7fdb6f4c-7tm5b" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.188441 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-config" (OuterVolumeSpecName: "config") pod "3860befc-70ca-4658-b967-b3e8351c9788" (UID: "3860befc-70ca-4658-b967-b3e8351c9788"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.190378 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbkxk" event={"ID":"4c12c954-d94a-404a-9263-a8371fdc12e4","Type":"ContainerDied","Data":"5f3f90c975a1130519cc70bf5fd2fe333ea91e0b04b54201e03255352eb01966"} Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.190422 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f3f90c975a1130519cc70bf5fd2fe333ea91e0b04b54201e03255352eb01966" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.190503 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbkxk" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.197459 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3860befc-70ca-4658-b967-b3e8351c9788" (UID: "3860befc-70ca-4658-b967-b3e8351c9788"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.198605 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3860befc-70ca-4658-b967-b3e8351c9788" (UID: "3860befc-70ca-4658-b967-b3e8351c9788"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.201155 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-config-data" (OuterVolumeSpecName: "config-data") pod "4c12c954-d94a-404a-9263-a8371fdc12e4" (UID: "4c12c954-d94a-404a-9263-a8371fdc12e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.202771 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" event={"ID":"3860befc-70ca-4658-b967-b3e8351c9788","Type":"ContainerDied","Data":"a0f2903ad0a5471038439b36c4b502fc1ed8e5184b388bf8079e0e6be436ffe4"} Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.202854 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.209557 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3860befc-70ca-4658-b967-b3e8351c9788" (UID: "3860befc-70ca-4658-b967-b3e8351c9788"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.209563 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68758c88fc-fgg5d" event={"ID":"17b02f27-1f55-4ecc-b7c2-f860720857d2","Type":"ContainerDied","Data":"41275cc6199fc83c0e792cc009bd5ffc308c1b0d085d6aecd90c0e137137fa40"} Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.209665 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68758c88fc-fgg5d" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.214571 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-976df9d9-dcx8g" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.215033 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-976df9d9-dcx8g" event={"ID":"17405477-0e11-40ac-b33d-b1b614827780","Type":"ContainerDied","Data":"ca97e40f9c7d0553e8703a3acd14d40194c3adf5c07ee4405472827ce402fd08"} Dec 03 14:10:32 crc kubenswrapper[4677]: E1203 14:10:32.216291 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-hgx4j" podUID="25625b10-c45e-40f0-b342-bdcdc3e79c9c" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.237727 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b02f27-1f55-4ecc-b7c2-f860720857d2-logs\") pod \"17b02f27-1f55-4ecc-b7c2-f860720857d2\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.237797 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-scripts\") pod \"17b02f27-1f55-4ecc-b7c2-f860720857d2\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.237838 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17b02f27-1f55-4ecc-b7c2-f860720857d2-horizon-secret-key\") pod \"17b02f27-1f55-4ecc-b7c2-f860720857d2\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238108 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djss7\" (UniqueName: \"kubernetes.io/projected/17b02f27-1f55-4ecc-b7c2-f860720857d2-kube-api-access-djss7\") pod \"17b02f27-1f55-4ecc-b7c2-f860720857d2\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238173 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-config-data\") pod \"17b02f27-1f55-4ecc-b7c2-f860720857d2\" (UID: \"17b02f27-1f55-4ecc-b7c2-f860720857d2\") " Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238627 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238646 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238851 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238862 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c12c954-d94a-404a-9263-a8371fdc12e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238873 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.238883 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3860befc-70ca-4658-b967-b3e8351c9788-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.240473 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17b02f27-1f55-4ecc-b7c2-f860720857d2-logs" (OuterVolumeSpecName: "logs") pod "17b02f27-1f55-4ecc-b7c2-f860720857d2" (UID: "17b02f27-1f55-4ecc-b7c2-f860720857d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.240590 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-scripts" (OuterVolumeSpecName: "scripts") pod "17b02f27-1f55-4ecc-b7c2-f860720857d2" (UID: "17b02f27-1f55-4ecc-b7c2-f860720857d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.244299 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-config-data" (OuterVolumeSpecName: "config-data") pod "17b02f27-1f55-4ecc-b7c2-f860720857d2" (UID: "17b02f27-1f55-4ecc-b7c2-f860720857d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.247152 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17b02f27-1f55-4ecc-b7c2-f860720857d2-kube-api-access-djss7" (OuterVolumeSpecName: "kube-api-access-djss7") pod "17b02f27-1f55-4ecc-b7c2-f860720857d2" (UID: "17b02f27-1f55-4ecc-b7c2-f860720857d2"). InnerVolumeSpecName "kube-api-access-djss7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.254167 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17b02f27-1f55-4ecc-b7c2-f860720857d2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "17b02f27-1f55-4ecc-b7c2-f860720857d2" (UID: "17b02f27-1f55-4ecc-b7c2-f860720857d2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.256234 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b7fdb6f4c-7tm5b"] Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.263851 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-b7fdb6f4c-7tm5b"] Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.319779 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-976df9d9-dcx8g"] Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.327581 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-976df9d9-dcx8g"] Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.340560 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djss7\" (UniqueName: \"kubernetes.io/projected/17b02f27-1f55-4ecc-b7c2-f860720857d2-kube-api-access-djss7\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.340603 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.340618 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b02f27-1f55-4ecc-b7c2-f860720857d2-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.340629 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17b02f27-1f55-4ecc-b7c2-f860720857d2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.340640 4677 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17b02f27-1f55-4ecc-b7c2-f860720857d2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.614135 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68758c88fc-fgg5d"] Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.621891 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68758c88fc-fgg5d"] Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.630131 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj"] Dec 03 14:10:32 crc kubenswrapper[4677]: I1203 14:10:32.638359 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj"] Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.266023 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c"] Dec 03 14:10:33 crc kubenswrapper[4677]: E1203 14:10:33.266449 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c12c954-d94a-404a-9263-a8371fdc12e4" containerName="glance-db-sync" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.266462 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c12c954-d94a-404a-9263-a8371fdc12e4" containerName="glance-db-sync" Dec 03 14:10:33 crc kubenswrapper[4677]: E1203 14:10:33.266479 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="init" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.266485 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="init" Dec 03 14:10:33 crc kubenswrapper[4677]: E1203 14:10:33.266497 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.266503 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.266668 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c12c954-d94a-404a-9263-a8371fdc12e4" containerName="glance-db-sync" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.266693 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.267680 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.283441 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c"] Dec 03 14:10:33 crc kubenswrapper[4677]: E1203 14:10:33.323019 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Dec 03 14:10:33 crc kubenswrapper[4677]: E1203 14:10:33.323075 4677 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Dec 03 14:10:33 crc kubenswrapper[4677]: E1203 14:10:33.327105 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.129.56.48:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6tctg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-njl82_openstack(0fd09b0c-1690-42e1-b97a-a18e7fefc813): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:10:33 crc kubenswrapper[4677]: E1203 14:10:33.328309 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-njl82" podUID="0fd09b0c-1690-42e1-b97a-a18e7fefc813" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.331055 4677 scope.go:117] "RemoveContainer" containerID="54bde3200fb30d3600edb18d2265f2ec1bd5e215d84a37a590b65bc099fa43a3" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.375333 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-config\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.375604 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.375710 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-svc\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.375810 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvfl9\" (UniqueName: \"kubernetes.io/projected/991ccfb3-a839-473d-9c0b-874b5ff7fe60-kube-api-access-pvfl9\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.376044 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.376170 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.478312 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-config\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.478355 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.478374 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-svc\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.478401 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvfl9\" (UniqueName: \"kubernetes.io/projected/991ccfb3-a839-473d-9c0b-874b5ff7fe60-kube-api-access-pvfl9\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.478505 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.478537 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.479507 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.479509 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.480385 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-svc\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.481522 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-config\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.482223 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.507852 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-72sks" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.509331 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvfl9\" (UniqueName: \"kubernetes.io/projected/991ccfb3-a839-473d-9c0b-874b5ff7fe60-kube-api-access-pvfl9\") pod \"dnsmasq-dns-6f9c8fbfbf-rvh4c\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.580619 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25vw9\" (UniqueName: \"kubernetes.io/projected/d88b54f3-fe9e-4247-8292-1f513f84fd33-kube-api-access-25vw9\") pod \"d88b54f3-fe9e-4247-8292-1f513f84fd33\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.580917 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-config\") pod \"d88b54f3-fe9e-4247-8292-1f513f84fd33\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.581181 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-combined-ca-bundle\") pod \"d88b54f3-fe9e-4247-8292-1f513f84fd33\" (UID: \"d88b54f3-fe9e-4247-8292-1f513f84fd33\") " Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.588832 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d88b54f3-fe9e-4247-8292-1f513f84fd33-kube-api-access-25vw9" (OuterVolumeSpecName: "kube-api-access-25vw9") pod "d88b54f3-fe9e-4247-8292-1f513f84fd33" (UID: "d88b54f3-fe9e-4247-8292-1f513f84fd33"). InnerVolumeSpecName "kube-api-access-25vw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.617449 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d88b54f3-fe9e-4247-8292-1f513f84fd33" (UID: "d88b54f3-fe9e-4247-8292-1f513f84fd33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.628350 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-config" (OuterVolumeSpecName: "config") pod "d88b54f3-fe9e-4247-8292-1f513f84fd33" (UID: "d88b54f3-fe9e-4247-8292-1f513f84fd33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.685019 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.685055 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25vw9\" (UniqueName: \"kubernetes.io/projected/d88b54f3-fe9e-4247-8292-1f513f84fd33-kube-api-access-25vw9\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.685068 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d88b54f3-fe9e-4247-8292-1f513f84fd33-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.804472 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.909806 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85b5c78d96-8262j"] Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.986915 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17405477-0e11-40ac-b33d-b1b614827780" path="/var/lib/kubelet/pods/17405477-0e11-40ac-b33d-b1b614827780/volumes" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.987369 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17b02f27-1f55-4ecc-b7c2-f860720857d2" path="/var/lib/kubelet/pods/17b02f27-1f55-4ecc-b7c2-f860720857d2/volumes" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.987716 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3860befc-70ca-4658-b967-b3e8351c9788" path="/var/lib/kubelet/pods/3860befc-70ca-4658-b967-b3e8351c9788/volumes" Dec 03 14:10:33 crc kubenswrapper[4677]: I1203 14:10:33.988499 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="852927ce-1113-484c-a27f-5e9f75c7f7f0" path="/var/lib/kubelet/pods/852927ce-1113-484c-a27f-5e9f75c7f7f0/volumes" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.015413 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c9d4fbd46-n2z95"] Dec 03 14:10:34 crc kubenswrapper[4677]: W1203 14:10:34.063362 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ec3794_7d5b_49dc_a957_bee4761d55f0.slice/crio-15153bdcffcb20da026bd3f7fcf706ce181873b5a84c6df8b05b4093603fe971 WatchSource:0}: Error finding container 15153bdcffcb20da026bd3f7fcf706ce181873b5a84c6df8b05b4093603fe971: Status 404 returned error can't find the container with id 15153bdcffcb20da026bd3f7fcf706ce181873b5a84c6df8b05b4093603fe971 Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.144784 4677 scope.go:117] "RemoveContainer" containerID="f754ecd4d121f270a28d322a0c1dfc8b5a8dc2cfc26968d7adfe3d3f99188ebd" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.196806 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:34 crc kubenswrapper[4677]: E1203 14:10:34.197227 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88b54f3-fe9e-4247-8292-1f513f84fd33" containerName="neutron-db-sync" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.197240 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88b54f3-fe9e-4247-8292-1f513f84fd33" containerName="neutron-db-sync" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.202524 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88b54f3-fe9e-4247-8292-1f513f84fd33" containerName="neutron-db-sync" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.203765 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.207672 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dl8zp" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.208602 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.208755 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.263232 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-72sks" event={"ID":"d88b54f3-fe9e-4247-8292-1f513f84fd33","Type":"ContainerDied","Data":"b4d5e022a662596ddae59eda08164f8be2af48f132d618e8479e36e0c8646a63"} Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.263536 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4d5e022a662596ddae59eda08164f8be2af48f132d618e8479e36e0c8646a63" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.263595 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-72sks" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.268828 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.274623 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b5c78d96-8262j" event={"ID":"c4ec3794-7d5b-49dc-a957-bee4761d55f0","Type":"ContainerStarted","Data":"15153bdcffcb20da026bd3f7fcf706ce181873b5a84c6df8b05b4093603fe971"} Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.278104 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c9d4fbd46-n2z95" event={"ID":"0749a528-b8b2-497f-bde8-f7b7765e8068","Type":"ContainerStarted","Data":"e830d3c3f9728cb8717bf4dd4b916610c6b0b9933cb94222d6d49e84851e125c"} Dec 03 14:10:34 crc kubenswrapper[4677]: E1203 14:10:34.288698 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-njl82" podUID="0fd09b0c-1690-42e1-b97a-a18e7fefc813" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.293710 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cbbb6d9d7-5z4fj" podUID="3860befc-70ca-4658-b967-b3e8351c9788" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.298092 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.298140 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-logs\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.298156 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.298176 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvd77\" (UniqueName: \"kubernetes.io/projected/c1176a9e-d925-485c-800d-47157987798e-kube-api-access-qvd77\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.298207 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.298293 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-scripts\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.298371 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-config-data\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.313215 4677 scope.go:117] "RemoveContainer" containerID="ae5ab29a0972e40bec375a1634edb74351d79e1fd5e1d4ec70437460e2232b5c" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.364144 4677 scope.go:117] "RemoveContainer" containerID="4f1344de56e500782cf8e8b798c387a14aa88f6e84cb7cd399436ae56ddce1d2" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.400224 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-scripts\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.400318 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-config-data\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.400339 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.400365 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-logs\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.400384 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.400399 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvd77\" (UniqueName: \"kubernetes.io/projected/c1176a9e-d925-485c-800d-47157987798e-kube-api-access-qvd77\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.400430 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.404400 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-logs\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.415069 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.415376 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.416413 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.418000 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-scripts\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.421824 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-config-data\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.444968 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvd77\" (UniqueName: \"kubernetes.io/projected/c1176a9e-d925-485c-800d-47157987798e-kube-api-access-qvd77\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.447754 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.449499 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.452854 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.463744 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.546735 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.596581 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.614103 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.614167 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.614196 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpt6t\" (UniqueName: \"kubernetes.io/projected/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-kube-api-access-gpt6t\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.614248 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.614312 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.614338 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.618189 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.660410 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ff56g"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.721210 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.721593 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.721618 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpt6t\" (UniqueName: \"kubernetes.io/projected/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-kube-api-access-gpt6t\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.721655 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.721707 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.721726 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.721747 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.722311 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.722554 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.731324 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.731657 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.744355 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.745537 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.759779 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpt6t\" (UniqueName: \"kubernetes.io/projected/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-kube-api-access-gpt6t\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.862153 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.921873 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.936014 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7857f7fffc-qb4s8"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.937551 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.939889 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.953278 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7857f7fffc-qb4s8"] Dec 03 14:10:34 crc kubenswrapper[4677]: I1203 14:10:34.988518 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c"] Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.036439 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-svc\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.036697 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-swift-storage-0\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.036734 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-config\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.036816 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-nb\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.036848 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-sb\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.036881 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkr8v\" (UniqueName: \"kubernetes.io/projected/fa7ee2cb-a257-4314-b745-788aa64ca1de-kube-api-access-zkr8v\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.123166 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.144585 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-nb\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.144701 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-sb\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.144797 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkr8v\" (UniqueName: \"kubernetes.io/projected/fa7ee2cb-a257-4314-b745-788aa64ca1de-kube-api-access-zkr8v\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.144856 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-svc\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.144905 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-swift-storage-0\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.144996 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-config\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.154224 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-nb\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.155757 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-sb\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.156323 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-config\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.157417 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-svc\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.167263 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-swift-storage-0\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.168905 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5574fd4bb8-fs46f"] Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.170447 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.183443 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9nljg" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.183490 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.183846 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.183983 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.232354 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5574fd4bb8-fs46f"] Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.308136 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerStarted","Data":"5f8248e38c43d4427d7c586ea1b85645e3cbc9ad237b8dd3d5ef2c546003a0bc"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.333240 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.346709 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f74nz" event={"ID":"24ec7657-9bc6-4953-9b8c-d07a62b7aff1","Type":"ContainerStarted","Data":"acb134cff354677e06caeafe5b47abfe39f4ca0e2990bea9a3e8de61c3fc2a5e"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.351700 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-combined-ca-bundle\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.352196 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb6j4\" (UniqueName: \"kubernetes.io/projected/1e5be0cc-a23f-4d36-a5a1-405d14814402-kube-api-access-nb6j4\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.352261 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-httpd-config\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.352365 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-ovndb-tls-certs\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.352439 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-config\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.355462 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c9d4fbd46-n2z95" event={"ID":"0749a528-b8b2-497f-bde8-f7b7765e8068","Type":"ContainerStarted","Data":"c6c156d4b68e1a727c0cdbb45a48bccccd10762ba4775c7f3e2f52b211e1423e"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.371357 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkr8v\" (UniqueName: \"kubernetes.io/projected/fa7ee2cb-a257-4314-b745-788aa64ca1de-kube-api-access-zkr8v\") pod \"dnsmasq-dns-7857f7fffc-qb4s8\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.424438 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.433473 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5edce1f1-9c11-4049-b286-a16939acb2e7","Type":"ContainerStarted","Data":"e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.440238 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"b40fca00-06a6-4109-bfc5-d67649e1a51c","Type":"ContainerStarted","Data":"d2490e1cb44cc90005875df405b48318257264bfbe866af6271cf7d98f141370"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.445150 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" event={"ID":"991ccfb3-a839-473d-9c0b-874b5ff7fe60","Type":"ContainerStarted","Data":"0a6e1633d3f0d3ebadcb6c15da0f5f1b119ebc34723c9017a822975569ce37ad"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.447461 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"1e9739b2-3472-4adb-9917-20b334a308cb","Type":"ContainerStarted","Data":"aff8e7061db29f457c9bb9f189981b88c5f32ba42157c0a64be8c4295f279880"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.448636 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ff56g" event={"ID":"47d54aae-4e2b-454f-a29d-9b252fd60179","Type":"ContainerStarted","Data":"0615795dfc8f674bf5695c0a05ede78c7bd85e0a9c93d47850f543abec9b262a"} Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.453715 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-combined-ca-bundle\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.453819 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb6j4\" (UniqueName: \"kubernetes.io/projected/1e5be0cc-a23f-4d36-a5a1-405d14814402-kube-api-access-nb6j4\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.453852 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-httpd-config\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.453997 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-ovndb-tls-certs\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.454083 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-config\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.472529 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-f74nz" podStartSLOduration=6.621958139 podStartE2EDuration="39.472508273s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="2025-12-03 14:09:58.856231321 +0000 UTC m=+1389.602563776" lastFinishedPulling="2025-12-03 14:10:31.706781455 +0000 UTC m=+1422.453113910" observedRunningTime="2025-12-03 14:10:35.43445762 +0000 UTC m=+1426.180790075" watchObservedRunningTime="2025-12-03 14:10:35.472508273 +0000 UTC m=+1426.218840738" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.479207 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-combined-ca-bundle\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.480090 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-httpd-config\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.480935 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=19.233572543 podStartE2EDuration="39.480921508s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="2025-12-03 14:09:58.364247848 +0000 UTC m=+1389.110580293" lastFinishedPulling="2025-12-03 14:10:18.611596803 +0000 UTC m=+1409.357929258" observedRunningTime="2025-12-03 14:10:35.454842151 +0000 UTC m=+1426.201174606" watchObservedRunningTime="2025-12-03 14:10:35.480921508 +0000 UTC m=+1426.227253973" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.492894 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb6j4\" (UniqueName: \"kubernetes.io/projected/1e5be0cc-a23f-4d36-a5a1-405d14814402-kube-api-access-nb6j4\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.495765 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-ovndb-tls-certs\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.498910 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-config\") pod \"neutron-5574fd4bb8-fs46f\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.531351 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=6.155323284 podStartE2EDuration="39.531331937s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="2025-12-03 14:09:58.336278923 +0000 UTC m=+1389.082611378" lastFinishedPulling="2025-12-03 14:10:31.712287576 +0000 UTC m=+1422.458620031" observedRunningTime="2025-12-03 14:10:35.490385111 +0000 UTC m=+1426.236717566" watchObservedRunningTime="2025-12-03 14:10:35.531331937 +0000 UTC m=+1426.277664392" Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.539003 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:35 crc kubenswrapper[4677]: I1203 14:10:35.544033 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.323483 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7857f7fffc-qb4s8"] Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.490337 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" event={"ID":"fa7ee2cb-a257-4314-b745-788aa64ca1de","Type":"ContainerStarted","Data":"f3e5abfef88e94510724c8ffa1fdff5fc5b011a0f51bca91069e2b058e7be20a"} Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.504256 4677 generic.go:334] "Generic (PLEG): container finished" podID="991ccfb3-a839-473d-9c0b-874b5ff7fe60" containerID="ad6f02a764c2cfc89542fdbbeabb8cbcfc5cabdca5d9d5df5b45438fab1ac5ab" exitCode=0 Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.504333 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" event={"ID":"991ccfb3-a839-473d-9c0b-874b5ff7fe60","Type":"ContainerDied","Data":"ad6f02a764c2cfc89542fdbbeabb8cbcfc5cabdca5d9d5df5b45438fab1ac5ab"} Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.529979 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c1176a9e-d925-485c-800d-47157987798e","Type":"ContainerStarted","Data":"cee5f2c94e8d33ba55b855178029bd44bc146dc875778498232d623c515d7e40"} Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.553700 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"1e9739b2-3472-4adb-9917-20b334a308cb","Type":"ContainerStarted","Data":"4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346"} Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.606062 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ff56g" event={"ID":"47d54aae-4e2b-454f-a29d-9b252fd60179","Type":"ContainerStarted","Data":"14b20e13076483ca1cbb724b3434ea4e9211378877f0effb2a1ece94cb676338"} Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.633485 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b5c78d96-8262j" event={"ID":"c4ec3794-7d5b-49dc-a957-bee4761d55f0","Type":"ContainerStarted","Data":"e6cccf2feedc2eccb7024eb91dee2e49a9b56d56ebf604895bef4e840ca87f49"} Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.650878 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c9d4fbd46-n2z95" event={"ID":"0749a528-b8b2-497f-bde8-f7b7765e8068","Type":"ContainerStarted","Data":"00073fc26bc8bb6894861e97accc9230cbe7c605934cacc112938af16f4d134f"} Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.655640 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.663470 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ff56g" podStartSLOduration=16.663449484 podStartE2EDuration="16.663449484s" podCreationTimestamp="2025-12-03 14:10:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:36.639647656 +0000 UTC m=+1427.385980121" watchObservedRunningTime="2025-12-03 14:10:36.663449484 +0000 UTC m=+1427.409781939" Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.692565 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c9d4fbd46-n2z95" podStartSLOduration=31.418596411 podStartE2EDuration="31.692534018s" podCreationTimestamp="2025-12-03 14:10:05 +0000 UTC" firstStartedPulling="2025-12-03 14:10:34.070542114 +0000 UTC m=+1424.816874569" lastFinishedPulling="2025-12-03 14:10:34.344479721 +0000 UTC m=+1425.090812176" observedRunningTime="2025-12-03 14:10:36.692451226 +0000 UTC m=+1427.438783691" watchObservedRunningTime="2025-12-03 14:10:36.692534018 +0000 UTC m=+1427.438866473" Dec 03 14:10:36 crc kubenswrapper[4677]: I1203 14:10:36.778749 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5574fd4bb8-fs46f"] Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.105290 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.105812 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.118976 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.182803 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.266169 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.270230 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.309447 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.316819 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-config\") pod \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.316858 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-sb\") pod \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.316891 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-nb\") pod \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.316921 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvfl9\" (UniqueName: \"kubernetes.io/projected/991ccfb3-a839-473d-9c0b-874b5ff7fe60-kube-api-access-pvfl9\") pod \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.317157 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-svc\") pod \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.317201 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-swift-storage-0\") pod \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\" (UID: \"991ccfb3-a839-473d-9c0b-874b5ff7fe60\") " Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.324893 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/991ccfb3-a839-473d-9c0b-874b5ff7fe60-kube-api-access-pvfl9" (OuterVolumeSpecName: "kube-api-access-pvfl9") pod "991ccfb3-a839-473d-9c0b-874b5ff7fe60" (UID: "991ccfb3-a839-473d-9c0b-874b5ff7fe60"). InnerVolumeSpecName "kube-api-access-pvfl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.350791 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "991ccfb3-a839-473d-9c0b-874b5ff7fe60" (UID: "991ccfb3-a839-473d-9c0b-874b5ff7fe60"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.368758 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-config" (OuterVolumeSpecName: "config") pod "991ccfb3-a839-473d-9c0b-874b5ff7fe60" (UID: "991ccfb3-a839-473d-9c0b-874b5ff7fe60"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.371736 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "991ccfb3-a839-473d-9c0b-874b5ff7fe60" (UID: "991ccfb3-a839-473d-9c0b-874b5ff7fe60"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.380635 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "991ccfb3-a839-473d-9c0b-874b5ff7fe60" (UID: "991ccfb3-a839-473d-9c0b-874b5ff7fe60"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.398033 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "991ccfb3-a839-473d-9c0b-874b5ff7fe60" (UID: "991ccfb3-a839-473d-9c0b-874b5ff7fe60"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.418333 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.419125 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.419149 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.419158 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.419167 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvfl9\" (UniqueName: \"kubernetes.io/projected/991ccfb3-a839-473d-9c0b-874b5ff7fe60-kube-api-access-pvfl9\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.419175 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.419183 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/991ccfb3-a839-473d-9c0b-874b5ff7fe60-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.822511 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"1e9739b2-3472-4adb-9917-20b334a308cb","Type":"ContainerStarted","Data":"1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.823154 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.825679 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce","Type":"ContainerStarted","Data":"6537c1f2e50b933501f939e74c4bdc323a751b70eb981070502ebdc1d64af871"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.826918 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b5c78d96-8262j" event={"ID":"c4ec3794-7d5b-49dc-a957-bee4761d55f0","Type":"ContainerStarted","Data":"8a66e68be6fff49b84696df5494a916693c024f8458f6735679bc7c8ba30309a"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.833270 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" event={"ID":"fa7ee2cb-a257-4314-b745-788aa64ca1de","Type":"ContainerStarted","Data":"1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.886467 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" event={"ID":"991ccfb3-a839-473d-9c0b-874b5ff7fe60","Type":"ContainerDied","Data":"0a6e1633d3f0d3ebadcb6c15da0f5f1b119ebc34723c9017a822975569ce37ad"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.886756 4677 scope.go:117] "RemoveContainer" containerID="ad6f02a764c2cfc89542fdbbeabb8cbcfc5cabdca5d9d5df5b45438fab1ac5ab" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.886891 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.905420 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=18.905405371 podStartE2EDuration="18.905405371s" podCreationTimestamp="2025-12-03 14:10:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:37.864698869 +0000 UTC m=+1428.611031314" watchObservedRunningTime="2025-12-03 14:10:37.905405371 +0000 UTC m=+1428.651737826" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.920043 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c1176a9e-d925-485c-800d-47157987798e","Type":"ContainerStarted","Data":"2c59932ef79b74926e762a98e520db79b944de4907aa7823c9bc4fe4db69190a"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.937573 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-85b5c78d96-8262j" podStartSLOduration=32.641648253 podStartE2EDuration="32.937555152s" podCreationTimestamp="2025-12-03 14:10:05 +0000 UTC" firstStartedPulling="2025-12-03 14:10:34.072492104 +0000 UTC m=+1424.818824559" lastFinishedPulling="2025-12-03 14:10:34.368399003 +0000 UTC m=+1425.114731458" observedRunningTime="2025-12-03 14:10:37.901400138 +0000 UTC m=+1428.647732593" watchObservedRunningTime="2025-12-03 14:10:37.937555152 +0000 UTC m=+1428.683887607" Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.955976 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5574fd4bb8-fs46f" event={"ID":"1e5be0cc-a23f-4d36-a5a1-405d14814402","Type":"ContainerStarted","Data":"2b1d3f58d578305185c5dc629c41dbf1e2307cf556f7b0097c814545841fbbfa"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.956011 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5574fd4bb8-fs46f" event={"ID":"1e5be0cc-a23f-4d36-a5a1-405d14814402","Type":"ContainerStarted","Data":"710ecce1ef0310635d23db5287532bc42741253c26e57418757aac09eb1e1647"} Dec 03 14:10:37 crc kubenswrapper[4677]: I1203 14:10:37.957231 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:38 crc kubenswrapper[4677]: I1203 14:10:38.067463 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Dec 03 14:10:38 crc kubenswrapper[4677]: I1203 14:10:38.125337 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:38 crc kubenswrapper[4677]: I1203 14:10:38.153773 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:10:38 crc kubenswrapper[4677]: I1203 14:10:38.175582 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:10:38 crc kubenswrapper[4677]: I1203 14:10:38.974790 4677 generic.go:334] "Generic (PLEG): container finished" podID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerID="1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d" exitCode=0 Dec 03 14:10:38 crc kubenswrapper[4677]: I1203 14:10:38.974886 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" event={"ID":"fa7ee2cb-a257-4314-b745-788aa64ca1de","Type":"ContainerDied","Data":"1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d"} Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.467378 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-55bcdc4f7c-89g9k"] Dec 03 14:10:39 crc kubenswrapper[4677]: E1203 14:10:39.467780 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="991ccfb3-a839-473d-9c0b-874b5ff7fe60" containerName="init" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.467797 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="991ccfb3-a839-473d-9c0b-874b5ff7fe60" containerName="init" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.467991 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="991ccfb3-a839-473d-9c0b-874b5ff7fe60" containerName="init" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.469365 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.471817 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.481606 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.484939 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55bcdc4f7c-89g9k"] Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.497771 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-public-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.497844 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lck\" (UniqueName: \"kubernetes.io/projected/1ea332c2-0df8-470e-8985-e68ef50ed410-kube-api-access-85lck\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.497900 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-config\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.497988 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-httpd-config\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.498042 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-combined-ca-bundle\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.498084 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-internal-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.498192 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-ovndb-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.515560 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.515751 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.600149 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-public-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.600247 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lck\" (UniqueName: \"kubernetes.io/projected/1ea332c2-0df8-470e-8985-e68ef50ed410-kube-api-access-85lck\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.600293 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-config\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.600344 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-httpd-config\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.600387 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-combined-ca-bundle\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.600416 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-internal-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.600491 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-ovndb-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.609306 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-combined-ca-bundle\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.609632 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-internal-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.617268 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-config\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.626334 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-httpd-config\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.630644 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lck\" (UniqueName: \"kubernetes.io/projected/1ea332c2-0df8-470e-8985-e68ef50ed410-kube-api-access-85lck\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.638769 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-ovndb-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.639530 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea332c2-0df8-470e-8985-e68ef50ed410-public-tls-certs\") pod \"neutron-55bcdc4f7c-89g9k\" (UID: \"1ea332c2-0df8-470e-8985-e68ef50ed410\") " pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:39 crc kubenswrapper[4677]: I1203 14:10:39.796677 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:40 crc kubenswrapper[4677]: I1203 14:10:40.013131 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:10:40 crc kubenswrapper[4677]: I1203 14:10:40.013578 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" containerID="cri-o://e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" gracePeriod=30 Dec 03 14:10:40 crc kubenswrapper[4677]: I1203 14:10:40.013748 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="b40fca00-06a6-4109-bfc5-d67649e1a51c" containerName="watcher-decision-engine" containerID="cri-o://d2490e1cb44cc90005875df405b48318257264bfbe866af6271cf7d98f141370" gracePeriod=30 Dec 03 14:10:40 crc kubenswrapper[4677]: I1203 14:10:40.556130 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/watcher-api-0" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:10:41 crc kubenswrapper[4677]: I1203 14:10:41.027622 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce","Type":"ContainerStarted","Data":"16c52dfe42196cd5989a78c6008b63ba3877cd55e3f3c6de25a4084a5d135c0d"} Dec 03 14:10:41 crc kubenswrapper[4677]: I1203 14:10:41.035823 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:10:41 crc kubenswrapper[4677]: I1203 14:10:41.036656 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5574fd4bb8-fs46f" event={"ID":"1e5be0cc-a23f-4d36-a5a1-405d14814402","Type":"ContainerStarted","Data":"4326918e7f551aea44b394c6bfbd72322962e712dde1ff6592bd906c64244143"} Dec 03 14:10:41 crc kubenswrapper[4677]: I1203 14:10:41.529265 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 14:10:42 crc kubenswrapper[4677]: E1203 14:10:42.109360 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:42 crc kubenswrapper[4677]: E1203 14:10:42.112446 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:42 crc kubenswrapper[4677]: E1203 14:10:42.114250 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:42 crc kubenswrapper[4677]: E1203 14:10:42.114288 4677 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:10:43 crc kubenswrapper[4677]: I1203 14:10:43.478362 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55bcdc4f7c-89g9k"] Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.067071 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c1176a9e-d925-485c-800d-47157987798e","Type":"ContainerStarted","Data":"a620970366abfdec8a0e0821e9c62eddea58e1e4bed8a14f79ef8647c5c81210"} Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.067145 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-log" containerID="cri-o://2c59932ef79b74926e762a98e520db79b944de4907aa7823c9bc4fe4db69190a" gracePeriod=30 Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.067284 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-httpd" containerID="cri-o://a620970366abfdec8a0e0821e9c62eddea58e1e4bed8a14f79ef8647c5c81210" gracePeriod=30 Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.070134 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" event={"ID":"fa7ee2cb-a257-4314-b745-788aa64ca1de","Type":"ContainerStarted","Data":"f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36"} Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.070251 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.088285 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.088268642 podStartE2EDuration="11.088268642s" podCreationTimestamp="2025-12-03 14:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:44.084473472 +0000 UTC m=+1434.830805927" watchObservedRunningTime="2025-12-03 14:10:44.088268642 +0000 UTC m=+1434.834601107" Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.119100 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" podStartSLOduration=10.119082178 podStartE2EDuration="10.119082178s" podCreationTimestamp="2025-12-03 14:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:44.107524936 +0000 UTC m=+1434.853857391" watchObservedRunningTime="2025-12-03 14:10:44.119082178 +0000 UTC m=+1434.865414623" Dec 03 14:10:44 crc kubenswrapper[4677]: I1203 14:10:44.136589 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5574fd4bb8-fs46f" podStartSLOduration=9.136569495 podStartE2EDuration="9.136569495s" podCreationTimestamp="2025-12-03 14:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:44.12489575 +0000 UTC m=+1434.871228225" watchObservedRunningTime="2025-12-03 14:10:44.136569495 +0000 UTC m=+1434.882901950" Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.096851 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1176a9e-d925-485c-800d-47157987798e" containerID="a620970366abfdec8a0e0821e9c62eddea58e1e4bed8a14f79ef8647c5c81210" exitCode=0 Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.097210 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1176a9e-d925-485c-800d-47157987798e" containerID="2c59932ef79b74926e762a98e520db79b944de4907aa7823c9bc4fe4db69190a" exitCode=143 Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.097007 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c1176a9e-d925-485c-800d-47157987798e","Type":"ContainerDied","Data":"a620970366abfdec8a0e0821e9c62eddea58e1e4bed8a14f79ef8647c5c81210"} Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.097315 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c1176a9e-d925-485c-800d-47157987798e","Type":"ContainerDied","Data":"2c59932ef79b74926e762a98e520db79b944de4907aa7823c9bc4fe4db69190a"} Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.101213 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce","Type":"ContainerStarted","Data":"72536bb0b4e9df5914be6919a46b9f82fc7a331f5bf9cd21f16b3fd10ecf8bb3"} Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.101353 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-log" containerID="cri-o://16c52dfe42196cd5989a78c6008b63ba3877cd55e3f3c6de25a4084a5d135c0d" gracePeriod=30 Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.101374 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-httpd" containerID="cri-o://72536bb0b4e9df5914be6919a46b9f82fc7a331f5bf9cd21f16b3fd10ecf8bb3" gracePeriod=30 Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.102650 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.105521 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5574fd4bb8-fs46f" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.870574 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.870904 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.993988 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.995068 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:45 crc kubenswrapper[4677]: I1203 14:10:45.999089 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c9d4fbd46-n2z95" podUID="0749a528-b8b2-497f-bde8-f7b7765e8068" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.158:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.158:8443: connect: connection refused" Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.111808 4677 generic.go:334] "Generic (PLEG): container finished" podID="47d54aae-4e2b-454f-a29d-9b252fd60179" containerID="14b20e13076483ca1cbb724b3434ea4e9211378877f0effb2a1ece94cb676338" exitCode=0 Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.111879 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ff56g" event={"ID":"47d54aae-4e2b-454f-a29d-9b252fd60179","Type":"ContainerDied","Data":"14b20e13076483ca1cbb724b3434ea4e9211378877f0effb2a1ece94cb676338"} Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.114445 4677 generic.go:334] "Generic (PLEG): container finished" podID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerID="72536bb0b4e9df5914be6919a46b9f82fc7a331f5bf9cd21f16b3fd10ecf8bb3" exitCode=0 Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.114469 4677 generic.go:334] "Generic (PLEG): container finished" podID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerID="16c52dfe42196cd5989a78c6008b63ba3877cd55e3f3c6de25a4084a5d135c0d" exitCode=143 Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.114507 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce","Type":"ContainerDied","Data":"72536bb0b4e9df5914be6919a46b9f82fc7a331f5bf9cd21f16b3fd10ecf8bb3"} Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.114541 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce","Type":"ContainerDied","Data":"16c52dfe42196cd5989a78c6008b63ba3877cd55e3f3c6de25a4084a5d135c0d"} Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.118489 4677 generic.go:334] "Generic (PLEG): container finished" podID="24ec7657-9bc6-4953-9b8c-d07a62b7aff1" containerID="acb134cff354677e06caeafe5b47abfe39f4ca0e2990bea9a3e8de61c3fc2a5e" exitCode=0 Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.119145 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f74nz" event={"ID":"24ec7657-9bc6-4953-9b8c-d07a62b7aff1","Type":"ContainerDied","Data":"acb134cff354677e06caeafe5b47abfe39f4ca0e2990bea9a3e8de61c3fc2a5e"} Dec 03 14:10:46 crc kubenswrapper[4677]: I1203 14:10:46.135921 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=13.1358976 podStartE2EDuration="13.1358976s" podCreationTimestamp="2025-12-03 14:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:45.144819112 +0000 UTC m=+1435.891151567" watchObservedRunningTime="2025-12-03 14:10:46.1358976 +0000 UTC m=+1436.882230055" Dec 03 14:10:46 crc kubenswrapper[4677]: W1203 14:10:46.694944 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ea332c2_0df8_470e_8985_e68ef50ed410.slice/crio-451830056f616c3377d18971e099c454354f1b309dea62a327c2727aa4bd77d9 WatchSource:0}: Error finding container 451830056f616c3377d18971e099c454354f1b309dea62a327c2727aa4bd77d9: Status 404 returned error can't find the container with id 451830056f616c3377d18971e099c454354f1b309dea62a327c2727aa4bd77d9 Dec 03 14:10:47 crc kubenswrapper[4677]: E1203 14:10:47.106337 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:47 crc kubenswrapper[4677]: E1203 14:10:47.110797 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:47 crc kubenswrapper[4677]: E1203 14:10:47.129608 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:47 crc kubenswrapper[4677]: E1203 14:10:47.129673 4677 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.143766 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55bcdc4f7c-89g9k" event={"ID":"1ea332c2-0df8-470e-8985-e68ef50ed410","Type":"ContainerStarted","Data":"0ece208c22d66a8c20f4c173443c62c4b12d48db87b1004e40f34ad2a8505a98"} Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.143837 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55bcdc4f7c-89g9k" event={"ID":"1ea332c2-0df8-470e-8985-e68ef50ed410","Type":"ContainerStarted","Data":"451830056f616c3377d18971e099c454354f1b309dea62a327c2727aa4bd77d9"} Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.156318 4677 generic.go:334] "Generic (PLEG): container finished" podID="b40fca00-06a6-4109-bfc5-d67649e1a51c" containerID="d2490e1cb44cc90005875df405b48318257264bfbe866af6271cf7d98f141370" exitCode=1 Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.156485 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"b40fca00-06a6-4109-bfc5-d67649e1a51c","Type":"ContainerDied","Data":"d2490e1cb44cc90005875df405b48318257264bfbe866af6271cf7d98f141370"} Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.161547 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hgx4j" event={"ID":"25625b10-c45e-40f0-b342-bdcdc3e79c9c","Type":"ContainerStarted","Data":"12faf5c4818df4effa9de71c604ec4af267f4cd5b4a81c1535a6a7c5920ed641"} Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.166700 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerStarted","Data":"1ec3750646ee7fe2cd08896eb5a41aea1e9aae3c83da20d8000b7fa9033fde29"} Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.192140 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-hgx4j" podStartSLOduration=3.2230944790000002 podStartE2EDuration="51.192123152s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="2025-12-03 14:09:58.946008768 +0000 UTC m=+1389.692341223" lastFinishedPulling="2025-12-03 14:10:46.915037441 +0000 UTC m=+1437.661369896" observedRunningTime="2025-12-03 14:10:47.189421142 +0000 UTC m=+1437.935753597" watchObservedRunningTime="2025-12-03 14:10:47.192123152 +0000 UTC m=+1437.938455607" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.322181 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.325410 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.402993 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.462735 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"c1176a9e-d925-485c-800d-47157987798e\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.462890 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvd77\" (UniqueName: \"kubernetes.io/projected/c1176a9e-d925-485c-800d-47157987798e-kube-api-access-qvd77\") pod \"c1176a9e-d925-485c-800d-47157987798e\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.462933 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-config-data\") pod \"b40fca00-06a6-4109-bfc5-d67649e1a51c\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463223 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-httpd-run\") pod \"c1176a9e-d925-485c-800d-47157987798e\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463274 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-scripts\") pod \"c1176a9e-d925-485c-800d-47157987798e\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463310 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40fca00-06a6-4109-bfc5-d67649e1a51c-logs\") pod \"b40fca00-06a6-4109-bfc5-d67649e1a51c\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463342 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-config-data\") pod \"c1176a9e-d925-485c-800d-47157987798e\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463384 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-combined-ca-bundle\") pod \"c1176a9e-d925-485c-800d-47157987798e\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463410 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-custom-prometheus-ca\") pod \"b40fca00-06a6-4109-bfc5-d67649e1a51c\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463441 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-httpd-run\") pod \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463463 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vffd9\" (UniqueName: \"kubernetes.io/projected/b40fca00-06a6-4109-bfc5-d67649e1a51c-kube-api-access-vffd9\") pod \"b40fca00-06a6-4109-bfc5-d67649e1a51c\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463490 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-combined-ca-bundle\") pod \"b40fca00-06a6-4109-bfc5-d67649e1a51c\" (UID: \"b40fca00-06a6-4109-bfc5-d67649e1a51c\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463513 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-logs\") pod \"c1176a9e-d925-485c-800d-47157987798e\" (UID: \"c1176a9e-d925-485c-800d-47157987798e\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463538 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463741 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40fca00-06a6-4109-bfc5-d67649e1a51c-logs" (OuterVolumeSpecName: "logs") pod "b40fca00-06a6-4109-bfc5-d67649e1a51c" (UID: "b40fca00-06a6-4109-bfc5-d67649e1a51c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.463881 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40fca00-06a6-4109-bfc5-d67649e1a51c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.464008 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c1176a9e-d925-485c-800d-47157987798e" (UID: "c1176a9e-d925-485c-800d-47157987798e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.466160 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-logs" (OuterVolumeSpecName: "logs") pod "c1176a9e-d925-485c-800d-47157987798e" (UID: "c1176a9e-d925-485c-800d-47157987798e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.466560 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" (UID: "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.466943 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1176a9e-d925-485c-800d-47157987798e-kube-api-access-qvd77" (OuterVolumeSpecName: "kube-api-access-qvd77") pod "c1176a9e-d925-485c-800d-47157987798e" (UID: "c1176a9e-d925-485c-800d-47157987798e"). InnerVolumeSpecName "kube-api-access-qvd77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.478320 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "c1176a9e-d925-485c-800d-47157987798e" (UID: "c1176a9e-d925-485c-800d-47157987798e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.478503 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" (UID: "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.478549 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-scripts" (OuterVolumeSpecName: "scripts") pod "c1176a9e-d925-485c-800d-47157987798e" (UID: "c1176a9e-d925-485c-800d-47157987798e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.498045 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40fca00-06a6-4109-bfc5-d67649e1a51c-kube-api-access-vffd9" (OuterVolumeSpecName: "kube-api-access-vffd9") pod "b40fca00-06a6-4109-bfc5-d67649e1a51c" (UID: "b40fca00-06a6-4109-bfc5-d67649e1a51c"). InnerVolumeSpecName "kube-api-access-vffd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.506405 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "b40fca00-06a6-4109-bfc5-d67649e1a51c" (UID: "b40fca00-06a6-4109-bfc5-d67649e1a51c"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.537532 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b40fca00-06a6-4109-bfc5-d67649e1a51c" (UID: "b40fca00-06a6-4109-bfc5-d67649e1a51c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.537819 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1176a9e-d925-485c-800d-47157987798e" (UID: "c1176a9e-d925-485c-800d-47157987798e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.546152 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-config-data" (OuterVolumeSpecName: "config-data") pod "b40fca00-06a6-4109-bfc5-d67649e1a51c" (UID: "b40fca00-06a6-4109-bfc5-d67649e1a51c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.564781 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-combined-ca-bundle\") pod \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.564822 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-scripts\") pod \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.564841 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-logs\") pod \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.564859 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-config-data\") pod \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.564892 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpt6t\" (UniqueName: \"kubernetes.io/projected/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-kube-api-access-gpt6t\") pod \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\" (UID: \"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565294 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565314 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565325 4677 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565338 4677 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565348 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vffd9\" (UniqueName: \"kubernetes.io/projected/b40fca00-06a6-4109-bfc5-d67649e1a51c-kube-api-access-vffd9\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565359 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565370 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565389 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565401 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565411 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvd77\" (UniqueName: \"kubernetes.io/projected/c1176a9e-d925-485c-800d-47157987798e-kube-api-access-qvd77\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565420 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40fca00-06a6-4109-bfc5-d67649e1a51c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.565428 4677 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c1176a9e-d925-485c-800d-47157987798e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.568344 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-logs" (OuterVolumeSpecName: "logs") pod "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" (UID: "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.572898 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-scripts" (OuterVolumeSpecName: "scripts") pod "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" (UID: "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.573587 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.580028 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-kube-api-access-gpt6t" (OuterVolumeSpecName: "kube-api-access-gpt6t") pod "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" (UID: "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce"). InnerVolumeSpecName "kube-api-access-gpt6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.619737 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" (UID: "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.652114 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-config-data" (OuterVolumeSpecName: "config-data") pod "c1176a9e-d925-485c-800d-47157987798e" (UID: "c1176a9e-d925-485c-800d-47157987798e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.668023 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-scripts\") pod \"47d54aae-4e2b-454f-a29d-9b252fd60179\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.668089 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbpw5\" (UniqueName: \"kubernetes.io/projected/47d54aae-4e2b-454f-a29d-9b252fd60179-kube-api-access-mbpw5\") pod \"47d54aae-4e2b-454f-a29d-9b252fd60179\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.668128 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-config-data\") pod \"47d54aae-4e2b-454f-a29d-9b252fd60179\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.668174 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-credential-keys\") pod \"47d54aae-4e2b-454f-a29d-9b252fd60179\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.668229 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-fernet-keys\") pod \"47d54aae-4e2b-454f-a29d-9b252fd60179\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.668271 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-combined-ca-bundle\") pod \"47d54aae-4e2b-454f-a29d-9b252fd60179\" (UID: \"47d54aae-4e2b-454f-a29d-9b252fd60179\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.669541 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1176a9e-d925-485c-800d-47157987798e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.669558 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.669568 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.669576 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.669587 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpt6t\" (UniqueName: \"kubernetes.io/projected/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-kube-api-access-gpt6t\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.673146 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d54aae-4e2b-454f-a29d-9b252fd60179-kube-api-access-mbpw5" (OuterVolumeSpecName: "kube-api-access-mbpw5") pod "47d54aae-4e2b-454f-a29d-9b252fd60179" (UID: "47d54aae-4e2b-454f-a29d-9b252fd60179"). InnerVolumeSpecName "kube-api-access-mbpw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.686151 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "47d54aae-4e2b-454f-a29d-9b252fd60179" (UID: "47d54aae-4e2b-454f-a29d-9b252fd60179"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.688134 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-scripts" (OuterVolumeSpecName: "scripts") pod "47d54aae-4e2b-454f-a29d-9b252fd60179" (UID: "47d54aae-4e2b-454f-a29d-9b252fd60179"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.692289 4677 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.692662 4677 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.702128 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "47d54aae-4e2b-454f-a29d-9b252fd60179" (UID: "47d54aae-4e2b-454f-a29d-9b252fd60179"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.713186 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-config-data" (OuterVolumeSpecName: "config-data") pod "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" (UID: "bb5ddc3b-348c-4488-b8ae-31f7a7f590ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.720691 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47d54aae-4e2b-454f-a29d-9b252fd60179" (UID: "47d54aae-4e2b-454f-a29d-9b252fd60179"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.742380 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-config-data" (OuterVolumeSpecName: "config-data") pod "47d54aae-4e2b-454f-a29d-9b252fd60179" (UID: "47d54aae-4e2b-454f-a29d-9b252fd60179"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771560 4677 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771593 4677 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771603 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771614 4677 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771622 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771630 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771638 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbpw5\" (UniqueName: \"kubernetes.io/projected/47d54aae-4e2b-454f-a29d-9b252fd60179-kube-api-access-mbpw5\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771647 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.771655 4677 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d54aae-4e2b-454f-a29d-9b252fd60179-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.922716 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f74nz" Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.984337 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-logs\") pod \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.984440 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-combined-ca-bundle\") pod \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.984603 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-config-data\") pod \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.984660 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxdr4\" (UniqueName: \"kubernetes.io/projected/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-kube-api-access-zxdr4\") pod \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.984709 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-scripts\") pod \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\" (UID: \"24ec7657-9bc6-4953-9b8c-d07a62b7aff1\") " Dec 03 14:10:47 crc kubenswrapper[4677]: I1203 14:10:47.988388 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-logs" (OuterVolumeSpecName: "logs") pod "24ec7657-9bc6-4953-9b8c-d07a62b7aff1" (UID: "24ec7657-9bc6-4953-9b8c-d07a62b7aff1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.020639 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-scripts" (OuterVolumeSpecName: "scripts") pod "24ec7657-9bc6-4953-9b8c-d07a62b7aff1" (UID: "24ec7657-9bc6-4953-9b8c-d07a62b7aff1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.066180 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-kube-api-access-zxdr4" (OuterVolumeSpecName: "kube-api-access-zxdr4") pod "24ec7657-9bc6-4953-9b8c-d07a62b7aff1" (UID: "24ec7657-9bc6-4953-9b8c-d07a62b7aff1"). InnerVolumeSpecName "kube-api-access-zxdr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.085101 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24ec7657-9bc6-4953-9b8c-d07a62b7aff1" (UID: "24ec7657-9bc6-4953-9b8c-d07a62b7aff1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.086362 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxdr4\" (UniqueName: \"kubernetes.io/projected/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-kube-api-access-zxdr4\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.086381 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.086389 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.086397 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.186083 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-config-data" (OuterVolumeSpecName: "config-data") pod "24ec7657-9bc6-4953-9b8c-d07a62b7aff1" (UID: "24ec7657-9bc6-4953-9b8c-d07a62b7aff1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.187549 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ec7657-9bc6-4953-9b8c-d07a62b7aff1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.233207 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bb5ddc3b-348c-4488-b8ae-31f7a7f590ce","Type":"ContainerDied","Data":"6537c1f2e50b933501f939e74c4bdc323a751b70eb981070502ebdc1d64af871"} Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.233259 4677 scope.go:117] "RemoveContainer" containerID="72536bb0b4e9df5914be6919a46b9f82fc7a331f5bf9cd21f16b3fd10ecf8bb3" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.233380 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.255606 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55bcdc4f7c-89g9k" event={"ID":"1ea332c2-0df8-470e-8985-e68ef50ed410","Type":"ContainerStarted","Data":"7a37a08b7fa3e5cb2788b610fb6628069ee6bbadeae22b00f4d90a8ed5b9cf6b"} Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.256372 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.262592 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f74nz" event={"ID":"24ec7657-9bc6-4953-9b8c-d07a62b7aff1","Type":"ContainerDied","Data":"e3f6ff5bbdc3c9f850810a53e188cdd38088026399f295d9025a5b3e90a8fa7a"} Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.262625 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3f6ff5bbdc3c9f850810a53e188cdd38088026399f295d9025a5b3e90a8fa7a" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.262677 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f74nz" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.279569 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"b40fca00-06a6-4109-bfc5-d67649e1a51c","Type":"ContainerDied","Data":"7bc24f74215cf17a781c27d148298785755852661db2aeee03fcab1f522273dd"} Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.279710 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.284742 4677 scope.go:117] "RemoveContainer" containerID="16c52dfe42196cd5989a78c6008b63ba3877cd55e3f3c6de25a4084a5d135c0d" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.287454 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c1176a9e-d925-485c-800d-47157987798e","Type":"ContainerDied","Data":"cee5f2c94e8d33ba55b855178029bd44bc146dc875778498232d623c515d7e40"} Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.287568 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.291087 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.303893 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9ccdcc6d4-gw5r7"] Dec 03 14:10:48 crc kubenswrapper[4677]: E1203 14:10:48.307680 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-log" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.307712 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-log" Dec 03 14:10:48 crc kubenswrapper[4677]: E1203 14:10:48.307730 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ec7657-9bc6-4953-9b8c-d07a62b7aff1" containerName="placement-db-sync" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.307739 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ec7657-9bc6-4953-9b8c-d07a62b7aff1" containerName="placement-db-sync" Dec 03 14:10:48 crc kubenswrapper[4677]: E1203 14:10:48.307774 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40fca00-06a6-4109-bfc5-d67649e1a51c" containerName="watcher-decision-engine" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.307786 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40fca00-06a6-4109-bfc5-d67649e1a51c" containerName="watcher-decision-engine" Dec 03 14:10:48 crc kubenswrapper[4677]: E1203 14:10:48.307800 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-log" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.307809 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-log" Dec 03 14:10:48 crc kubenswrapper[4677]: E1203 14:10:48.307825 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-httpd" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.307834 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-httpd" Dec 03 14:10:48 crc kubenswrapper[4677]: E1203 14:10:48.307854 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-httpd" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.307863 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-httpd" Dec 03 14:10:48 crc kubenswrapper[4677]: E1203 14:10:48.307879 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d54aae-4e2b-454f-a29d-9b252fd60179" containerName="keystone-bootstrap" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.307887 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d54aae-4e2b-454f-a29d-9b252fd60179" containerName="keystone-bootstrap" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.308266 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-httpd" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.308290 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ec7657-9bc6-4953-9b8c-d07a62b7aff1" containerName="placement-db-sync" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.308319 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40fca00-06a6-4109-bfc5-d67649e1a51c" containerName="watcher-decision-engine" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.308328 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-httpd" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.308339 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1176a9e-d925-485c-800d-47157987798e" containerName="glance-log" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.308353 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" containerName="glance-log" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.308362 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d54aae-4e2b-454f-a29d-9b252fd60179" containerName="keystone-bootstrap" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.309477 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.313683 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.324812 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9ccdcc6d4-gw5r7"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.324861 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ff56g" event={"ID":"47d54aae-4e2b-454f-a29d-9b252fd60179","Type":"ContainerDied","Data":"0615795dfc8f674bf5695c0a05ede78c7bd85e0a9c93d47850f543abec9b262a"} Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.324885 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0615795dfc8f674bf5695c0a05ede78c7bd85e0a9c93d47850f543abec9b262a" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.324867 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ff56g" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.334939 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.335189 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.340448 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-67bf5cb8b6-4lwks"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.342467 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.367007 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qrxb8" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.367727 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.367863 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.368014 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.368135 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.373044 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.375033 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.383695 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.386037 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.386304 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.386537 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dl8zp" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.415696 4677 scope.go:117] "RemoveContainer" containerID="d2490e1cb44cc90005875df405b48318257264bfbe866af6271cf7d98f141370" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.462827 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67bf5cb8b6-4lwks"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.478009 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-55bcdc4f7c-89g9k" podStartSLOduration=9.477987254 podStartE2EDuration="9.477987254s" podCreationTimestamp="2025-12-03 14:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:48.342387445 +0000 UTC m=+1439.088719910" watchObservedRunningTime="2025-12-03 14:10:48.477987254 +0000 UTC m=+1439.224319709" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.490270 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.492256 4677 scope.go:117] "RemoveContainer" containerID="a620970366abfdec8a0e0821e9c62eddea58e1e4bed8a14f79ef8647c5c81210" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.496284 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-credential-keys\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.499792 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzdkh\" (UniqueName: \"kubernetes.io/projected/85993ef9-4908-479b-88bb-1a0d12b832e0-kube-api-access-nzdkh\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.500203 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-public-tls-certs\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.500299 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.500393 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-combined-ca-bundle\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.500509 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr25r\" (UniqueName: \"kubernetes.io/projected/372640de-08f9-4020-9ccf-294678119877-kube-api-access-pr25r\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.500679 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.500770 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503056 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-combined-ca-bundle\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503163 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btzns\" (UniqueName: \"kubernetes.io/projected/232a4115-1312-485b-8df6-cad4097e7b6f-kube-api-access-btzns\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503243 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-config-data\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503406 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372640de-08f9-4020-9ccf-294678119877-logs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503485 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-fernet-keys\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503543 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-internal-tls-certs\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503630 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-internal-tls-certs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.503695 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-public-tls-certs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.504028 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.504112 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-scripts\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.504235 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.504319 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-config-data\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.504392 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.504492 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-scripts\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.507441 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.506553 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.516084 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.529002 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.530350 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.536379 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.539599 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.550132 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.559161 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.574656 4677 scope.go:117] "RemoveContainer" containerID="2c59932ef79b74926e762a98e520db79b944de4907aa7823c9bc4fe4db69190a" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.575708 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.577670 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.580328 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.580793 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.589933 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.608808 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.608849 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-credential-keys\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.608872 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzdkh\" (UniqueName: \"kubernetes.io/projected/85993ef9-4908-479b-88bb-1a0d12b832e0-kube-api-access-nzdkh\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.608926 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-public-tls-certs\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.608946 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.608976 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-combined-ca-bundle\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609002 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr25r\" (UniqueName: \"kubernetes.io/projected/372640de-08f9-4020-9ccf-294678119877-kube-api-access-pr25r\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609039 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609062 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609081 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-combined-ca-bundle\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609100 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btzns\" (UniqueName: \"kubernetes.io/projected/232a4115-1312-485b-8df6-cad4097e7b6f-kube-api-access-btzns\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609147 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-config-data\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609171 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372640de-08f9-4020-9ccf-294678119877-logs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609188 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-fernet-keys\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609221 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-internal-tls-certs\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609245 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-internal-tls-certs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609261 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-public-tls-certs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609281 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609305 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-scripts\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609341 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609362 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-config-data\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609377 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.609390 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-scripts\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.612627 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.612757 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.614804 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.622818 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372640de-08f9-4020-9ccf-294678119877-logs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.623721 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.628228 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-scripts\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.632544 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-credential-keys\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.634401 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-public-tls-certs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.638680 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.650969 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-fernet-keys\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.653151 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.655588 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-internal-tls-certs\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.656601 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-config-data\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.674565 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-internal-tls-certs\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.703589 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-public-tls-certs\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.703787 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-combined-ca-bundle\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.706181 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-scripts\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.706886 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85993ef9-4908-479b-88bb-1a0d12b832e0-config-data\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.707924 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btzns\" (UniqueName: \"kubernetes.io/projected/232a4115-1312-485b-8df6-cad4097e7b6f-kube-api-access-btzns\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.710255 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.711374 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372640de-08f9-4020-9ccf-294678119877-combined-ca-bundle\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.712646 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.712820 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f93ff7e-f6ab-4c00-8284-70a4354e576a-logs\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.712847 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.712988 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.713692 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.714019 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.714283 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.714405 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-config-data\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.714543 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.714668 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhtgl\" (UniqueName: \"kubernetes.io/projected/6945799b-1144-45b8-be28-8de8ad7b257e-kube-api-access-mhtgl\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.714820 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-logs\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.714980 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cfjh\" (UniqueName: \"kubernetes.io/projected/0f93ff7e-f6ab-4c00-8284-70a4354e576a-kube-api-access-9cfjh\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.715075 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.715178 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzdkh\" (UniqueName: \"kubernetes.io/projected/85993ef9-4908-479b-88bb-1a0d12b832e0-kube-api-access-nzdkh\") pod \"keystone-9ccdcc6d4-gw5r7\" (UID: \"85993ef9-4908-479b-88bb-1a0d12b832e0\") " pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.731219 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr25r\" (UniqueName: \"kubernetes.io/projected/372640de-08f9-4020-9ccf-294678119877-kube-api-access-pr25r\") pod \"placement-67bf5cb8b6-4lwks\" (UID: \"372640de-08f9-4020-9ccf-294678119877\") " pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.767190 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.816921 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817035 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-config-data\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817075 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817096 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhtgl\" (UniqueName: \"kubernetes.io/projected/6945799b-1144-45b8-be28-8de8ad7b257e-kube-api-access-mhtgl\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817145 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-logs\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817178 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cfjh\" (UniqueName: \"kubernetes.io/projected/0f93ff7e-f6ab-4c00-8284-70a4354e576a-kube-api-access-9cfjh\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817194 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817245 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817270 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f93ff7e-f6ab-4c00-8284-70a4354e576a-logs\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817286 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817316 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817340 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.817360 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.824657 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-logs\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.824664 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.825003 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.825549 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f93ff7e-f6ab-4c00-8284-70a4354e576a-logs\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.836844 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.844678 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.845564 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.845584 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.847518 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.848436 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-config-data\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.856760 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.864150 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhtgl\" (UniqueName: \"kubernetes.io/projected/6945799b-1144-45b8-be28-8de8ad7b257e-kube-api-access-mhtgl\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.868259 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cfjh\" (UniqueName: \"kubernetes.io/projected/0f93ff7e-f6ab-4c00-8284-70a4354e576a-kube-api-access-9cfjh\") pod \"watcher-decision-engine-0\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.870889 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.936574 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.985731 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:48 crc kubenswrapper[4677]: I1203 14:10:48.995998 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:49 crc kubenswrapper[4677]: I1203 14:10:49.042585 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:49 crc kubenswrapper[4677]: I1203 14:10:49.159560 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:49 crc kubenswrapper[4677]: I1203 14:10:49.515524 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Dec 03 14:10:49 crc kubenswrapper[4677]: I1203 14:10:49.530990 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 14:10:49 crc kubenswrapper[4677]: I1203 14:10:49.581024 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9ccdcc6d4-gw5r7"] Dec 03 14:10:49 crc kubenswrapper[4677]: W1203 14:10:49.649373 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85993ef9_4908_479b_88bb_1a0d12b832e0.slice/crio-f5b232ca421c74cd82b369af05b0840df8979c851d525b4a0ab70a027aee406f WatchSource:0}: Error finding container f5b232ca421c74cd82b369af05b0840df8979c851d525b4a0ab70a027aee406f: Status 404 returned error can't find the container with id f5b232ca421c74cd82b369af05b0840df8979c851d525b4a0ab70a027aee406f Dec 03 14:10:49 crc kubenswrapper[4677]: I1203 14:10:49.733828 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67bf5cb8b6-4lwks"] Dec 03 14:10:49 crc kubenswrapper[4677]: I1203 14:10:49.831160 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.145289 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40fca00-06a6-4109-bfc5-d67649e1a51c" path="/var/lib/kubelet/pods/b40fca00-06a6-4109-bfc5-d67649e1a51c/volumes" Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.145919 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb5ddc3b-348c-4488-b8ae-31f7a7f590ce" path="/var/lib/kubelet/pods/bb5ddc3b-348c-4488-b8ae-31f7a7f590ce/volumes" Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.146774 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1176a9e-d925-485c-800d-47157987798e" path="/var/lib/kubelet/pods/c1176a9e-d925-485c-800d-47157987798e/volumes" Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.160426 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.223592 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.427551 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.439516 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67bf5cb8b6-4lwks" event={"ID":"372640de-08f9-4020-9ccf-294678119877","Type":"ContainerStarted","Data":"dce67134398367cb198d5e3119bbd536580bd97be1fd45d76274cd62c161c1e9"} Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.452624 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6945799b-1144-45b8-be28-8de8ad7b257e","Type":"ContainerStarted","Data":"46d8fa12adb8d86245b5f61f8e2d04381a0bbc48cf74a64fccfaf1a5ff5e3985"} Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.457791 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9ccdcc6d4-gw5r7" event={"ID":"85993ef9-4908-479b-88bb-1a0d12b832e0","Type":"ContainerStarted","Data":"f5b232ca421c74cd82b369af05b0840df8979c851d525b4a0ab70a027aee406f"} Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.461202 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"232a4115-1312-485b-8df6-cad4097e7b6f","Type":"ContainerStarted","Data":"6e3e15a73c88cd2288469c9158915a94a4ee573aa0a7b2b017e23a0b655ca53f"} Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.466723 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerStarted","Data":"52ff4c1ad22782bc249cca16979fb7c08be743778a30945ce2c47217347856ff"} Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.559651 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcc496dd5-87k8s"] Dec 03 14:10:50 crc kubenswrapper[4677]: I1203 14:10:50.559938 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" podUID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerName="dnsmasq-dns" containerID="cri-o://485c37947f3b1ab828005255c2341513bc22935bc837f4ede0426514259fadd1" gracePeriod=10 Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.488490 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerStarted","Data":"fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf"} Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.513652 4677 generic.go:334] "Generic (PLEG): container finished" podID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerID="485c37947f3b1ab828005255c2341513bc22935bc837f4ede0426514259fadd1" exitCode=0 Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.513714 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" event={"ID":"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37","Type":"ContainerDied","Data":"485c37947f3b1ab828005255c2341513bc22935bc837f4ede0426514259fadd1"} Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.527113 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67bf5cb8b6-4lwks" event={"ID":"372640de-08f9-4020-9ccf-294678119877","Type":"ContainerStarted","Data":"fe1f561756f1bb68541ead366d9168666e345c8a754570ac6be936b2662a4ce7"} Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.532168 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=3.532148524 podStartE2EDuration="3.532148524s" podCreationTimestamp="2025-12-03 14:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:51.520283514 +0000 UTC m=+1442.266615969" watchObservedRunningTime="2025-12-03 14:10:51.532148524 +0000 UTC m=+1442.278480979" Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.533701 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-njl82" event={"ID":"0fd09b0c-1690-42e1-b97a-a18e7fefc813","Type":"ContainerStarted","Data":"64425cc924e22eaa6e7d652322613611f220ab3d76f8ce2539e1804f9ab5321a"} Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.563924 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-njl82" podStartSLOduration=5.721788089 podStartE2EDuration="55.563900465s" podCreationTimestamp="2025-12-03 14:09:56 +0000 UTC" firstStartedPulling="2025-12-03 14:09:58.293976821 +0000 UTC m=+1389.040309276" lastFinishedPulling="2025-12-03 14:10:48.136089197 +0000 UTC m=+1438.882421652" observedRunningTime="2025-12-03 14:10:51.549388245 +0000 UTC m=+1442.295720700" watchObservedRunningTime="2025-12-03 14:10:51.563900465 +0000 UTC m=+1442.310232930" Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.577543 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9ccdcc6d4-gw5r7" event={"ID":"85993ef9-4908-479b-88bb-1a0d12b832e0","Type":"ContainerStarted","Data":"413c36a66e9d7e6d2d75c232bfb19101284d5c18aaaaf1ab703422e05b5c38e9"} Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.577920 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.599910 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-9ccdcc6d4-gw5r7" podStartSLOduration=3.599892907 podStartE2EDuration="3.599892907s" podCreationTimestamp="2025-12-03 14:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:51.5984594 +0000 UTC m=+1442.344791875" watchObservedRunningTime="2025-12-03 14:10:51.599892907 +0000 UTC m=+1442.346225362" Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.968544 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.981731 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-svc\") pod \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.981788 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77nq8\" (UniqueName: \"kubernetes.io/projected/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-kube-api-access-77nq8\") pod \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.981919 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-config\") pod \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.981978 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-swift-storage-0\") pod \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.982067 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-sb\") pod \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " Dec 03 14:10:51 crc kubenswrapper[4677]: I1203 14:10:51.982122 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-nb\") pod \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\" (UID: \"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37\") " Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.020155 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-kube-api-access-77nq8" (OuterVolumeSpecName: "kube-api-access-77nq8") pod "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" (UID: "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37"). InnerVolumeSpecName "kube-api-access-77nq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.087647 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77nq8\" (UniqueName: \"kubernetes.io/projected/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-kube-api-access-77nq8\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:52 crc kubenswrapper[4677]: E1203 14:10:52.133480 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:52 crc kubenswrapper[4677]: E1203 14:10:52.182210 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:52 crc kubenswrapper[4677]: E1203 14:10:52.187767 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:52 crc kubenswrapper[4677]: E1203 14:10:52.187811 4677 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.225864 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" (UID: "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.297622 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.369610 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" (UID: "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.381458 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" (UID: "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.388100 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-config" (OuterVolumeSpecName: "config") pod "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" (UID: "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.394534 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" (UID: "953a83f6-7c2f-4fd5-b3f7-0430d4c00b37"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.400281 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.400317 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.400358 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.400373 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.628847 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6945799b-1144-45b8-be28-8de8ad7b257e","Type":"ContainerStarted","Data":"e86632b6cc6cc51b64f89d50ed680cd0451cc04ee86ae40f18ab50470fc6ffef"} Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.637005 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"232a4115-1312-485b-8df6-cad4097e7b6f","Type":"ContainerStarted","Data":"76bf2f570e3bbbcf5bc48c476c42d98548b1d6d49fcb95f4d041d75dbed406f3"} Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.656004 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" event={"ID":"953a83f6-7c2f-4fd5-b3f7-0430d4c00b37","Type":"ContainerDied","Data":"cd637071b474d2843176fabf5f6e40a3770e98e4a725cbeb4b0cab40cf44e3ed"} Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.656477 4677 scope.go:117] "RemoveContainer" containerID="485c37947f3b1ab828005255c2341513bc22935bc837f4ede0426514259fadd1" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.656634 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc496dd5-87k8s" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.681005 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67bf5cb8b6-4lwks" event={"ID":"372640de-08f9-4020-9ccf-294678119877","Type":"ContainerStarted","Data":"46fa07b9d84e168bffa889efe0ffa16c07ee564c74b8c8cb3301b93fb9864890"} Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.681506 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.681528 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.729192 4677 scope.go:117] "RemoveContainer" containerID="900ccc4aa90c89810a4522a43cd33ff556c623eccea11cc243aae76dc4da172f" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.778853 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-67bf5cb8b6-4lwks" podStartSLOduration=4.778831301 podStartE2EDuration="4.778831301s" podCreationTimestamp="2025-12-03 14:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:52.706072617 +0000 UTC m=+1443.452405172" watchObservedRunningTime="2025-12-03 14:10:52.778831301 +0000 UTC m=+1443.525163766" Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.805989 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcc496dd5-87k8s"] Dec 03 14:10:52 crc kubenswrapper[4677]: I1203 14:10:52.814593 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dcc496dd5-87k8s"] Dec 03 14:10:53 crc kubenswrapper[4677]: I1203 14:10:53.716032 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6945799b-1144-45b8-be28-8de8ad7b257e","Type":"ContainerStarted","Data":"10245f6819d6abec12abcada82177e8e22fed9a10dafe05c55136c7f886838f9"} Dec 03 14:10:53 crc kubenswrapper[4677]: I1203 14:10:53.742144 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"232a4115-1312-485b-8df6-cad4097e7b6f","Type":"ContainerStarted","Data":"6510391c92397ab7f26b1567ed8a627f449c9ce0cae708a46a17278bbb069e05"} Dec 03 14:10:53 crc kubenswrapper[4677]: I1203 14:10:53.767136 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.767115886 podStartE2EDuration="5.767115886s" podCreationTimestamp="2025-12-03 14:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:53.741613078 +0000 UTC m=+1444.487945533" watchObservedRunningTime="2025-12-03 14:10:53.767115886 +0000 UTC m=+1444.513448341" Dec 03 14:10:53 crc kubenswrapper[4677]: I1203 14:10:53.782850 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.782825446 podStartE2EDuration="5.782825446s" podCreationTimestamp="2025-12-03 14:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:10:53.775521666 +0000 UTC m=+1444.521854121" watchObservedRunningTime="2025-12-03 14:10:53.782825446 +0000 UTC m=+1444.529157901" Dec 03 14:10:53 crc kubenswrapper[4677]: I1203 14:10:53.991364 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" path="/var/lib/kubelet/pods/953a83f6-7c2f-4fd5-b3f7-0430d4c00b37/volumes" Dec 03 14:10:54 crc kubenswrapper[4677]: I1203 14:10:54.151089 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:54 crc kubenswrapper[4677]: I1203 14:10:54.151351 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api-log" containerID="cri-o://4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346" gracePeriod=30 Dec 03 14:10:54 crc kubenswrapper[4677]: I1203 14:10:54.151489 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api" containerID="cri-o://1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d" gracePeriod=30 Dec 03 14:10:54 crc kubenswrapper[4677]: I1203 14:10:54.773867 4677 generic.go:334] "Generic (PLEG): container finished" podID="1e9739b2-3472-4adb-9917-20b334a308cb" containerID="4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346" exitCode=143 Dec 03 14:10:54 crc kubenswrapper[4677]: I1203 14:10:54.773974 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"1e9739b2-3472-4adb-9917-20b334a308cb","Type":"ContainerDied","Data":"4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346"} Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.014768 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9322/\": read tcp 10.217.0.2:45022->10.217.0.159:9322: read: connection reset by peer" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.014824 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.159:9322/\": read tcp 10.217.0.2:45032->10.217.0.159:9322: read: connection reset by peer" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.502511 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.513968 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-combined-ca-bundle\") pod \"1e9739b2-3472-4adb-9917-20b334a308cb\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.514064 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e9739b2-3472-4adb-9917-20b334a308cb-logs\") pod \"1e9739b2-3472-4adb-9917-20b334a308cb\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.514120 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-config-data\") pod \"1e9739b2-3472-4adb-9917-20b334a308cb\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.514281 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k97j\" (UniqueName: \"kubernetes.io/projected/1e9739b2-3472-4adb-9917-20b334a308cb-kube-api-access-9k97j\") pod \"1e9739b2-3472-4adb-9917-20b334a308cb\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.514309 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-custom-prometheus-ca\") pod \"1e9739b2-3472-4adb-9917-20b334a308cb\" (UID: \"1e9739b2-3472-4adb-9917-20b334a308cb\") " Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.514515 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e9739b2-3472-4adb-9917-20b334a308cb-logs" (OuterVolumeSpecName: "logs") pod "1e9739b2-3472-4adb-9917-20b334a308cb" (UID: "1e9739b2-3472-4adb-9917-20b334a308cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.514812 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e9739b2-3472-4adb-9917-20b334a308cb-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.527551 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e9739b2-3472-4adb-9917-20b334a308cb-kube-api-access-9k97j" (OuterVolumeSpecName: "kube-api-access-9k97j") pod "1e9739b2-3472-4adb-9917-20b334a308cb" (UID: "1e9739b2-3472-4adb-9917-20b334a308cb"). InnerVolumeSpecName "kube-api-access-9k97j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.559899 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e9739b2-3472-4adb-9917-20b334a308cb" (UID: "1e9739b2-3472-4adb-9917-20b334a308cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.560001 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "1e9739b2-3472-4adb-9917-20b334a308cb" (UID: "1e9739b2-3472-4adb-9917-20b334a308cb"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.617328 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k97j\" (UniqueName: \"kubernetes.io/projected/1e9739b2-3472-4adb-9917-20b334a308cb-kube-api-access-9k97j\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.617369 4677 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.617382 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.637639 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-config-data" (OuterVolumeSpecName: "config-data") pod "1e9739b2-3472-4adb-9917-20b334a308cb" (UID: "1e9739b2-3472-4adb-9917-20b334a308cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.719058 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e9739b2-3472-4adb-9917-20b334a308cb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.793592 4677 generic.go:334] "Generic (PLEG): container finished" podID="25625b10-c45e-40f0-b342-bdcdc3e79c9c" containerID="12faf5c4818df4effa9de71c604ec4af267f4cd5b4a81c1535a6a7c5920ed641" exitCode=0 Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.793706 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hgx4j" event={"ID":"25625b10-c45e-40f0-b342-bdcdc3e79c9c","Type":"ContainerDied","Data":"12faf5c4818df4effa9de71c604ec4af267f4cd5b4a81c1535a6a7c5920ed641"} Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.797796 4677 generic.go:334] "Generic (PLEG): container finished" podID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerID="fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf" exitCode=1 Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.797971 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerDied","Data":"fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf"} Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.800920 4677 generic.go:334] "Generic (PLEG): container finished" podID="1e9739b2-3472-4adb-9917-20b334a308cb" containerID="1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d" exitCode=0 Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.800988 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"1e9739b2-3472-4adb-9917-20b334a308cb","Type":"ContainerDied","Data":"1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d"} Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.801069 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"1e9739b2-3472-4adb-9917-20b334a308cb","Type":"ContainerDied","Data":"aff8e7061db29f457c9bb9f189981b88c5f32ba42157c0a64be8c4295f279880"} Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.801129 4677 scope.go:117] "RemoveContainer" containerID="1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.801671 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.802372 4677 scope.go:117] "RemoveContainer" containerID="fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.903144 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.930121 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961016 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:55 crc kubenswrapper[4677]: E1203 14:10:55.961491 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961506 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api" Dec 03 14:10:55 crc kubenswrapper[4677]: E1203 14:10:55.961516 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerName="init" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961522 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerName="init" Dec 03 14:10:55 crc kubenswrapper[4677]: E1203 14:10:55.961555 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api-log" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961562 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api-log" Dec 03 14:10:55 crc kubenswrapper[4677]: E1203 14:10:55.961589 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerName="dnsmasq-dns" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961595 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerName="dnsmasq-dns" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961758 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api-log" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961772 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="953a83f6-7c2f-4fd5-b3f7-0430d4c00b37" containerName="dnsmasq-dns" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.961786 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" containerName="watcher-api" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.962848 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.966851 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.967053 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Dec 03 14:10:55 crc kubenswrapper[4677]: I1203 14:10:55.967100 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.011078 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e9739b2-3472-4adb-9917-20b334a308cb" path="/var/lib/kubelet/pods/1e9739b2-3472-4adb-9917-20b334a308cb/volumes" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.012169 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.027831 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.027886 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.027959 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j2cv\" (UniqueName: \"kubernetes.io/projected/12d8e83b-2cf0-4731-9c86-6a75651386b8-kube-api-access-8j2cv\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.027983 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d8e83b-2cf0-4731-9c86-6a75651386b8-logs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.028020 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-config-data\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.028099 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.028118 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-public-tls-certs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.131120 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.131203 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.131271 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j2cv\" (UniqueName: \"kubernetes.io/projected/12d8e83b-2cf0-4731-9c86-6a75651386b8-kube-api-access-8j2cv\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.131305 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d8e83b-2cf0-4731-9c86-6a75651386b8-logs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.131363 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-config-data\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.131472 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.131504 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-public-tls-certs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.132022 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d8e83b-2cf0-4731-9c86-6a75651386b8-logs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.141059 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-config-data\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.144694 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.145155 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-public-tls-certs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.145164 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.146515 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d8e83b-2cf0-4731-9c86-6a75651386b8-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.153580 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j2cv\" (UniqueName: \"kubernetes.io/projected/12d8e83b-2cf0-4731-9c86-6a75651386b8-kube-api-access-8j2cv\") pod \"watcher-api-0\" (UID: \"12d8e83b-2cf0-4731-9c86-6a75651386b8\") " pod="openstack/watcher-api-0" Dec 03 14:10:56 crc kubenswrapper[4677]: I1203 14:10:56.315641 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 14:10:57 crc kubenswrapper[4677]: E1203 14:10:57.107308 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:57 crc kubenswrapper[4677]: E1203 14:10:57.109456 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:57 crc kubenswrapper[4677]: E1203 14:10:57.110728 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:10:57 crc kubenswrapper[4677]: E1203 14:10:57.110801 4677 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:10:57 crc kubenswrapper[4677]: I1203 14:10:57.751764 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:57 crc kubenswrapper[4677]: I1203 14:10:57.966858 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:58 crc kubenswrapper[4677]: I1203 14:10:58.849421 4677 generic.go:334] "Generic (PLEG): container finished" podID="0fd09b0c-1690-42e1-b97a-a18e7fefc813" containerID="64425cc924e22eaa6e7d652322613611f220ab3d76f8ce2539e1804f9ab5321a" exitCode=0 Dec 03 14:10:58 crc kubenswrapper[4677]: I1203 14:10:58.849476 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-njl82" event={"ID":"0fd09b0c-1690-42e1-b97a-a18e7fefc813","Type":"ContainerDied","Data":"64425cc924e22eaa6e7d652322613611f220ab3d76f8ce2539e1804f9ab5321a"} Dec 03 14:10:58 crc kubenswrapper[4677]: I1203 14:10:58.937340 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:10:58 crc kubenswrapper[4677]: I1203 14:10:58.937413 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:10:58 crc kubenswrapper[4677]: I1203 14:10:58.972385 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.002547 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.043208 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.043259 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.077703 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.082884 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.161108 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.161154 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.549106 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.782759 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-c9d4fbd46-n2z95" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.920928 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85b5c78d96-8262j"] Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.923346 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.923669 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85b5c78d96-8262j" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon-log" containerID="cri-o://e6cccf2feedc2eccb7024eb91dee2e49a9b56d56ebf604895bef4e840ca87f49" gracePeriod=30 Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.924618 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85b5c78d96-8262j" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon" containerID="cri-o://8a66e68be6fff49b84696df5494a916693c024f8458f6735679bc7c8ba30309a" gracePeriod=30 Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.926531 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.935248 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:10:59 crc kubenswrapper[4677]: I1203 14:10:59.936612 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:11:00 crc kubenswrapper[4677]: E1203 14:11:00.056806 4677 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/85c181d4918682bbdadcb39741b4d7cf9ae147615ca885bc023b1804a6ace059/diff" to get inode usage: stat /var/lib/containers/storage/overlay/85c181d4918682bbdadcb39741b4d7cf9ae147615ca885bc023b1804a6ace059/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-dcc496dd5-87k8s_953a83f6-7c2f-4fd5-b3f7-0430d4c00b37/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-dcc496dd5-87k8s_953a83f6-7c2f-4fd5-b3f7-0430d4c00b37/dnsmasq-dns/0.log: no such file or directory Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.679787 4677 scope.go:117] "RemoveContainer" containerID="4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.777672 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.781350 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-njl82" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.856807 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tctg\" (UniqueName: \"kubernetes.io/projected/0fd09b0c-1690-42e1-b97a-a18e7fefc813-kube-api-access-6tctg\") pod \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.856849 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-db-sync-config-data\") pod \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.856881 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-config-data\") pod \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.856907 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-combined-ca-bundle\") pod \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.856975 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-db-sync-config-data\") pod \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.857018 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fd09b0c-1690-42e1-b97a-a18e7fefc813-etc-machine-id\") pod \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.857034 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb7l9\" (UniqueName: \"kubernetes.io/projected/25625b10-c45e-40f0-b342-bdcdc3e79c9c-kube-api-access-pb7l9\") pod \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\" (UID: \"25625b10-c45e-40f0-b342-bdcdc3e79c9c\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.857060 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-combined-ca-bundle\") pod \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.857077 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-scripts\") pod \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\" (UID: \"0fd09b0c-1690-42e1-b97a-a18e7fefc813\") " Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.858452 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fd09b0c-1690-42e1-b97a-a18e7fefc813-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0fd09b0c-1690-42e1-b97a-a18e7fefc813" (UID: "0fd09b0c-1690-42e1-b97a-a18e7fefc813"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.866964 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25625b10-c45e-40f0-b342-bdcdc3e79c9c-kube-api-access-pb7l9" (OuterVolumeSpecName: "kube-api-access-pb7l9") pod "25625b10-c45e-40f0-b342-bdcdc3e79c9c" (UID: "25625b10-c45e-40f0-b342-bdcdc3e79c9c"). InnerVolumeSpecName "kube-api-access-pb7l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.870586 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-scripts" (OuterVolumeSpecName: "scripts") pod "0fd09b0c-1690-42e1-b97a-a18e7fefc813" (UID: "0fd09b0c-1690-42e1-b97a-a18e7fefc813"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.870637 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0fd09b0c-1690-42e1-b97a-a18e7fefc813" (UID: "0fd09b0c-1690-42e1-b97a-a18e7fefc813"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.881420 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fd09b0c-1690-42e1-b97a-a18e7fefc813-kube-api-access-6tctg" (OuterVolumeSpecName: "kube-api-access-6tctg") pod "0fd09b0c-1690-42e1-b97a-a18e7fefc813" (UID: "0fd09b0c-1690-42e1-b97a-a18e7fefc813"). InnerVolumeSpecName "kube-api-access-6tctg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.883739 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "25625b10-c45e-40f0-b342-bdcdc3e79c9c" (UID: "25625b10-c45e-40f0-b342-bdcdc3e79c9c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.898091 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fd09b0c-1690-42e1-b97a-a18e7fefc813" (UID: "0fd09b0c-1690-42e1-b97a-a18e7fefc813"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.904489 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25625b10-c45e-40f0-b342-bdcdc3e79c9c" (UID: "25625b10-c45e-40f0-b342-bdcdc3e79c9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.941042 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-njl82" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.941077 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-njl82" event={"ID":"0fd09b0c-1690-42e1-b97a-a18e7fefc813","Type":"ContainerDied","Data":"f99039f40832f2e19ff5f6dfbade0ff5509753b5d47c31912f7113c43996b123"} Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.941121 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99039f40832f2e19ff5f6dfbade0ff5509753b5d47c31912f7113c43996b123" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.944561 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hgx4j" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.944828 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hgx4j" event={"ID":"25625b10-c45e-40f0-b342-bdcdc3e79c9c","Type":"ContainerDied","Data":"0e8086cca4d17a62b570dc85b637b50f072e46f99feb791e5c32cd1c611f2cfc"} Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.944915 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e8086cca4d17a62b570dc85b637b50f072e46f99feb791e5c32cd1c611f2cfc" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958749 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958819 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tctg\" (UniqueName: \"kubernetes.io/projected/0fd09b0c-1690-42e1-b97a-a18e7fefc813-kube-api-access-6tctg\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958832 4677 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958841 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958850 4677 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25625b10-c45e-40f0-b342-bdcdc3e79c9c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958858 4677 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fd09b0c-1690-42e1-b97a-a18e7fefc813-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958866 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb7l9\" (UniqueName: \"kubernetes.io/projected/25625b10-c45e-40f0-b342-bdcdc3e79c9c-kube-api-access-pb7l9\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:00 crc kubenswrapper[4677]: I1203 14:11:00.958874 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.113074 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-config-data" (OuterVolumeSpecName: "config-data") pod "0fd09b0c-1690-42e1-b97a-a18e7fefc813" (UID: "0fd09b0c-1690-42e1-b97a-a18e7fefc813"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.164571 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fd09b0c-1690-42e1-b97a-a18e7fefc813-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.188741 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:01 crc kubenswrapper[4677]: E1203 14:11:01.189127 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25625b10-c45e-40f0-b342-bdcdc3e79c9c" containerName="barbican-db-sync" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.189144 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="25625b10-c45e-40f0-b342-bdcdc3e79c9c" containerName="barbican-db-sync" Dec 03 14:11:01 crc kubenswrapper[4677]: E1203 14:11:01.189160 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fd09b0c-1690-42e1-b97a-a18e7fefc813" containerName="cinder-db-sync" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.189166 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fd09b0c-1690-42e1-b97a-a18e7fefc813" containerName="cinder-db-sync" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.189335 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fd09b0c-1690-42e1-b97a-a18e7fefc813" containerName="cinder-db-sync" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.189357 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="25625b10-c45e-40f0-b342-bdcdc3e79c9c" containerName="barbican-db-sync" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.190301 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.198183 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.240015 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.266111 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-scripts\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.266189 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.266239 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.266289 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.266315 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.266361 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5dsl\" (UniqueName: \"kubernetes.io/projected/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-kube-api-access-t5dsl\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.289014 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57f57f74bc-885jf"] Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.290667 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.302515 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57f57f74bc-885jf"] Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.368747 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-svc\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.368799 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-scripts\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.368854 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-config\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.368885 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.368929 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.368995 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.369016 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.369033 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-nb\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.369076 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-swift-storage-0\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.369099 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5dsl\" (UniqueName: \"kubernetes.io/projected/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-kube-api-access-t5dsl\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.369134 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-sb\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.369161 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csm6h\" (UniqueName: \"kubernetes.io/projected/4fe15275-4e76-46cc-9979-87af4f8a5407-kube-api-access-csm6h\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.370362 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.377923 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.382577 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.387002 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.393411 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-scripts\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.400133 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.402181 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.406449 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.414603 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5dsl\" (UniqueName: \"kubernetes.io/projected/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-kube-api-access-t5dsl\") pod \"cinder-scheduler-0\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.431023 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.470966 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-scripts\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471037 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471059 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d5f483c-fe0a-448f-be69-5c6bedf0e235-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471095 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-nb\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471111 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471154 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-swift-storage-0\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471179 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d5f483c-fe0a-448f-be69-5c6bedf0e235-logs\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471216 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-sb\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471240 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csm6h\" (UniqueName: \"kubernetes.io/projected/4fe15275-4e76-46cc-9979-87af4f8a5407-kube-api-access-csm6h\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471268 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-svc\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471300 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471319 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b47k9\" (UniqueName: \"kubernetes.io/projected/0d5f483c-fe0a-448f-be69-5c6bedf0e235-kube-api-access-b47k9\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471341 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-config\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471855 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-nb\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.471986 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-config\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.472490 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-swift-storage-0\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.472560 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-svc\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.473120 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-sb\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.489025 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csm6h\" (UniqueName: \"kubernetes.io/projected/4fe15275-4e76-46cc-9979-87af4f8a5407-kube-api-access-csm6h\") pod \"dnsmasq-dns-57f57f74bc-885jf\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.548078 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.573288 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.573333 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b47k9\" (UniqueName: \"kubernetes.io/projected/0d5f483c-fe0a-448f-be69-5c6bedf0e235-kube-api-access-b47k9\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.573386 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-scripts\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.573431 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.573453 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d5f483c-fe0a-448f-be69-5c6bedf0e235-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.573486 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.573531 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d5f483c-fe0a-448f-be69-5c6bedf0e235-logs\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.575239 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d5f483c-fe0a-448f-be69-5c6bedf0e235-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.575872 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d5f483c-fe0a-448f-be69-5c6bedf0e235-logs\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.578307 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-scripts\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.580613 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.580833 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.581099 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.594531 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b47k9\" (UniqueName: \"kubernetes.io/projected/0d5f483c-fe0a-448f-be69-5c6bedf0e235-kube-api-access-b47k9\") pod \"cinder-api-0\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.736341 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.753893 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.984054 4677 generic.go:334] "Generic (PLEG): container finished" podID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerID="8a66e68be6fff49b84696df5494a916693c024f8458f6735679bc7c8ba30309a" exitCode=0 Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.984157 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.984172 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.984722 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:11:01 crc kubenswrapper[4677]: I1203 14:11:01.984740 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.005673 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b5c78d96-8262j" event={"ID":"c4ec3794-7d5b-49dc-a957-bee4761d55f0","Type":"ContainerDied","Data":"8a66e68be6fff49b84696df5494a916693c024f8458f6735679bc7c8ba30309a"} Dec 03 14:11:02 crc kubenswrapper[4677]: E1203 14:11:02.108514 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.119891 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-64fbb7dcd5-9hc8s"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.121431 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: E1203 14:11:02.126453 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.126770 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.126839 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d2zwm" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.127084 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 14:11:02 crc kubenswrapper[4677]: E1203 14:11:02.141423 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:11:02 crc kubenswrapper[4677]: E1203 14:11:02.141505 4677 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.162605 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64fbb7dcd5-9hc8s"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.182977 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-67bc86b468-qwl7k"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.189231 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.195255 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.225605 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-combined-ca-bundle\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.225681 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt6t2\" (UniqueName: \"kubernetes.io/projected/e5960f5e-97d6-45c7-b597-b7ae61478585-kube-api-access-jt6t2\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.225720 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-config-data\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.225796 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5960f5e-97d6-45c7-b597-b7ae61478585-logs\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.225883 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-config-data-custom\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.293787 4677 scope.go:117] "RemoveContainer" containerID="1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d" Dec 03 14:11:02 crc kubenswrapper[4677]: E1203 14:11:02.298506 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d\": container with ID starting with 1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d not found: ID does not exist" containerID="1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.298583 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d"} err="failed to get container status \"1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d\": rpc error: code = NotFound desc = could not find container \"1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d\": container with ID starting with 1e95acd8f3624734e7e9f6b5bd519386be7f4f3e6024dbd8631d37e96b3e962d not found: ID does not exist" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.298629 4677 scope.go:117] "RemoveContainer" containerID="4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346" Dec 03 14:11:02 crc kubenswrapper[4677]: E1203 14:11:02.319108 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346\": container with ID starting with 4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346 not found: ID does not exist" containerID="4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.319374 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346"} err="failed to get container status \"4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346\": rpc error: code = NotFound desc = could not find container \"4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346\": container with ID starting with 4a5763325a65e736ba5556940f58f28a20d49dbdf048a25adcc4458108e9d346 not found: ID does not exist" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.328834 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-combined-ca-bundle\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329023 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69z4f\" (UniqueName: \"kubernetes.io/projected/9f174682-8786-4f0e-8ce3-f72c5560bde6-kube-api-access-69z4f\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329130 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5960f5e-97d6-45c7-b597-b7ae61478585-logs\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329474 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-config-data-custom\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329554 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-config-data-custom\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329619 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-combined-ca-bundle\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329640 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f174682-8786-4f0e-8ce3-f72c5560bde6-logs\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329661 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-config-data\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329710 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt6t2\" (UniqueName: \"kubernetes.io/projected/e5960f5e-97d6-45c7-b597-b7ae61478585-kube-api-access-jt6t2\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.329755 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-config-data\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.331583 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5960f5e-97d6-45c7-b597-b7ae61478585-logs\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.343671 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-combined-ca-bundle\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.345528 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-config-data\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.379776 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5960f5e-97d6-45c7-b597-b7ae61478585-config-data-custom\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.395936 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67bc86b468-qwl7k"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.416698 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt6t2\" (UniqueName: \"kubernetes.io/projected/e5960f5e-97d6-45c7-b597-b7ae61478585-kube-api-access-jt6t2\") pod \"barbican-worker-64fbb7dcd5-9hc8s\" (UID: \"e5960f5e-97d6-45c7-b597-b7ae61478585\") " pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.435125 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-config-data-custom\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.435395 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f174682-8786-4f0e-8ce3-f72c5560bde6-logs\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.435472 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-config-data\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.435623 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-combined-ca-bundle\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.435710 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69z4f\" (UniqueName: \"kubernetes.io/projected/9f174682-8786-4f0e-8ce3-f72c5560bde6-kube-api-access-69z4f\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.441970 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f174682-8786-4f0e-8ce3-f72c5560bde6-logs\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.453440 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-config-data\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.453871 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-config-data-custom\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.476824 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57f57f74bc-885jf"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.488888 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f174682-8786-4f0e-8ce3-f72c5560bde6-combined-ca-bundle\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.506387 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f955c7f4c-hfdt9"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.508612 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69z4f\" (UniqueName: \"kubernetes.io/projected/9f174682-8786-4f0e-8ce3-f72c5560bde6-kube-api-access-69z4f\") pod \"barbican-keystone-listener-67bc86b468-qwl7k\" (UID: \"9f174682-8786-4f0e-8ce3-f72c5560bde6\") " pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.515313 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.516217 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f955c7f4c-hfdt9"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.640283 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.642083 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-config\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.642131 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-swift-storage-0\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.642348 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-nb\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.642374 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt8df\" (UniqueName: \"kubernetes.io/projected/8e1c647c-16bc-4421-9bc8-7395212da3fe-kube-api-access-jt8df\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.642438 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-sb\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.642470 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-svc\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.674244 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5cf99b568-bhfkp"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.675900 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.690866 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.700210 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cf99b568-bhfkp"] Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.743893 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745626 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-svc\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745705 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-config\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745726 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-swift-storage-0\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745772 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-combined-ca-bundle\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745816 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745851 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvw5j\" (UniqueName: \"kubernetes.io/projected/6c52f0ed-484d-44cf-85a3-17732f779171-kube-api-access-bvw5j\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745880 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt8df\" (UniqueName: \"kubernetes.io/projected/8e1c647c-16bc-4421-9bc8-7395212da3fe-kube-api-access-jt8df\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745897 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-nb\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745917 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c52f0ed-484d-44cf-85a3-17732f779171-logs\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745933 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data-custom\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.745984 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-sb\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.759741 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-sb\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.760324 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-svc\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.765971 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-nb\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.802555 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-swift-storage-0\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.802701 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-config\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.842863 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt8df\" (UniqueName: \"kubernetes.io/projected/8e1c647c-16bc-4421-9bc8-7395212da3fe-kube-api-access-jt8df\") pod \"dnsmasq-dns-5f955c7f4c-hfdt9\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.875223 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-combined-ca-bundle\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.875347 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.875442 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvw5j\" (UniqueName: \"kubernetes.io/projected/6c52f0ed-484d-44cf-85a3-17732f779171-kube-api-access-bvw5j\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.875513 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c52f0ed-484d-44cf-85a3-17732f779171-logs\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.875551 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data-custom\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.884928 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c52f0ed-484d-44cf-85a3-17732f779171-logs\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.891304 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.892479 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data-custom\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.894875 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.895319 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-combined-ca-bundle\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.941609 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvw5j\" (UniqueName: \"kubernetes.io/projected/6c52f0ed-484d-44cf-85a3-17732f779171-kube-api-access-bvw5j\") pod \"barbican-api-5cf99b568-bhfkp\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:02 crc kubenswrapper[4677]: I1203 14:11:02.979754 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.138369 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.388333 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.446266 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.528719 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57f57f74bc-885jf"] Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.842335 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.843480 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.894699 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:11:03 crc kubenswrapper[4677]: I1203 14:11:03.894800 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.041598 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.047710 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f57f74bc-885jf" event={"ID":"4fe15275-4e76-46cc-9979-87af4f8a5407","Type":"ContainerStarted","Data":"f1295fd1b6ed54877a65e8c7c7114521ef3320900525061f111302cba63589e0"} Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.054072 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerStarted","Data":"2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0"} Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.055744 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f","Type":"ContainerStarted","Data":"a9762b43b134307683fdf5fe77343fee6a501c27cc95e4b82bc3d93cdad77824"} Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.058982 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerStarted","Data":"baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1"} Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.061342 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d5f483c-fe0a-448f-be69-5c6bedf0e235","Type":"ContainerStarted","Data":"c8cea9e4f23cc2907255a9e1929a3c9249c9b5bd840231ea01c334295c477dc3"} Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.078822 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"12d8e83b-2cf0-4731-9c86-6a75651386b8","Type":"ContainerStarted","Data":"012269dcf6e92e4c7484f2b84e056dedbfcafcb98daa7afb56dd5685ff69bd40"} Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.204542 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67bc86b468-qwl7k"] Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.247973 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f955c7f4c-hfdt9"] Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.302314 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64fbb7dcd5-9hc8s"] Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.461515 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cf99b568-bhfkp"] Dec 03 14:11:04 crc kubenswrapper[4677]: E1203 14:11:04.465476 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" Dec 03 14:11:04 crc kubenswrapper[4677]: I1203 14:11:04.734140 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.121333 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cf99b568-bhfkp" event={"ID":"6c52f0ed-484d-44cf-85a3-17732f779171","Type":"ContainerStarted","Data":"3232cdfd108c4e0352b82b175a6602a022312fd143ea2399bae6a912e93b7348"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.131272 4677 generic.go:334] "Generic (PLEG): container finished" podID="4fe15275-4e76-46cc-9979-87af4f8a5407" containerID="a051da98bd2120337b9c4f8c0b0125f55447dd4eb7f0818e30439e09704c26fd" exitCode=0 Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.131338 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f57f74bc-885jf" event={"ID":"4fe15275-4e76-46cc-9979-87af4f8a5407","Type":"ContainerDied","Data":"a051da98bd2120337b9c4f8c0b0125f55447dd4eb7f0818e30439e09704c26fd"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.174573 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d5f483c-fe0a-448f-be69-5c6bedf0e235","Type":"ContainerStarted","Data":"cef027ba74898040c24a79f67c8c3f4c331596839de6a2032407493d873047b6"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.183188 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" event={"ID":"9f174682-8786-4f0e-8ce3-f72c5560bde6","Type":"ContainerStarted","Data":"142163c40b4867652b3ab1fae6fe10f6c3ce2e3c8944a1f5a05fb925023a2efa"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.203900 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"12d8e83b-2cf0-4731-9c86-6a75651386b8","Type":"ContainerStarted","Data":"aea57eb96f629da28540817e5b274c75bcd34000e9bab79932746c0b11404ec9"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.246218 4677 generic.go:334] "Generic (PLEG): container finished" podID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerID="f730497aa99e9a0c54fcfbb2cb7a0fb6913270a37dbdb39e9acb7ad5edadc3df" exitCode=0 Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.246326 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" event={"ID":"8e1c647c-16bc-4421-9bc8-7395212da3fe","Type":"ContainerDied","Data":"f730497aa99e9a0c54fcfbb2cb7a0fb6913270a37dbdb39e9acb7ad5edadc3df"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.246356 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" event={"ID":"8e1c647c-16bc-4421-9bc8-7395212da3fe","Type":"ContainerStarted","Data":"60157941364cb6f27e072665c5c08ddb24c7252112c2775762c7cec9938e3138"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.317320 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="ceilometer-notification-agent" containerID="cri-o://5f8248e38c43d4427d7c586ea1b85645e3cbc9ad237b8dd3d5ef2c546003a0bc" gracePeriod=30 Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.317726 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" event={"ID":"e5960f5e-97d6-45c7-b597-b7ae61478585","Type":"ContainerStarted","Data":"490b4ee036234661d89b262ca94fd581ccf403aae8065b6b8a6ef3e28fcb32c5"} Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.318087 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="sg-core" containerID="cri-o://1ec3750646ee7fe2cd08896eb5a41aea1e9aae3c83da20d8000b7fa9033fde29" gracePeriod=30 Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.318230 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="proxy-httpd" containerID="cri-o://baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1" gracePeriod=30 Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.318104 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.644783 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:11:05 crc kubenswrapper[4677]: I1203 14:11:05.871471 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85b5c78d96-8262j" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.034188 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.110155 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csm6h\" (UniqueName: \"kubernetes.io/projected/4fe15275-4e76-46cc-9979-87af4f8a5407-kube-api-access-csm6h\") pod \"4fe15275-4e76-46cc-9979-87af4f8a5407\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.110513 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-swift-storage-0\") pod \"4fe15275-4e76-46cc-9979-87af4f8a5407\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.110544 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-svc\") pod \"4fe15275-4e76-46cc-9979-87af4f8a5407\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.111143 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-nb\") pod \"4fe15275-4e76-46cc-9979-87af4f8a5407\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.111173 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-sb\") pod \"4fe15275-4e76-46cc-9979-87af4f8a5407\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.111552 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-config\") pod \"4fe15275-4e76-46cc-9979-87af4f8a5407\" (UID: \"4fe15275-4e76-46cc-9979-87af4f8a5407\") " Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.136392 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fe15275-4e76-46cc-9979-87af4f8a5407-kube-api-access-csm6h" (OuterVolumeSpecName: "kube-api-access-csm6h") pod "4fe15275-4e76-46cc-9979-87af4f8a5407" (UID: "4fe15275-4e76-46cc-9979-87af4f8a5407"). InnerVolumeSpecName "kube-api-access-csm6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.216576 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csm6h\" (UniqueName: \"kubernetes.io/projected/4fe15275-4e76-46cc-9979-87af4f8a5407-kube-api-access-csm6h\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.329461 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"12d8e83b-2cf0-4731-9c86-6a75651386b8","Type":"ContainerStarted","Data":"14e24ed405bcc08f8d78f873f32c93bf346533cab9686009bbdcfa7e50fc8af5"} Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.332011 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.338271 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cf99b568-bhfkp" event={"ID":"6c52f0ed-484d-44cf-85a3-17732f779171","Type":"ContainerStarted","Data":"738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525"} Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.346316 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f57f74bc-885jf" event={"ID":"4fe15275-4e76-46cc-9979-87af4f8a5407","Type":"ContainerDied","Data":"f1295fd1b6ed54877a65e8c7c7114521ef3320900525061f111302cba63589e0"} Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.346359 4677 scope.go:117] "RemoveContainer" containerID="a051da98bd2120337b9c4f8c0b0125f55447dd4eb7f0818e30439e09704c26fd" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.346483 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f57f74bc-885jf" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.348972 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=11.348941023 podStartE2EDuration="11.348941023s" podCreationTimestamp="2025-12-03 14:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:06.347035823 +0000 UTC m=+1457.093368288" watchObservedRunningTime="2025-12-03 14:11:06.348941023 +0000 UTC m=+1457.095273468" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.357815 4677 generic.go:334] "Generic (PLEG): container finished" podID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerID="baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1" exitCode=0 Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.357849 4677 generic.go:334] "Generic (PLEG): container finished" podID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerID="1ec3750646ee7fe2cd08896eb5a41aea1e9aae3c83da20d8000b7fa9033fde29" exitCode=2 Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.357902 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerDied","Data":"baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1"} Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.357993 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerDied","Data":"1ec3750646ee7fe2cd08896eb5a41aea1e9aae3c83da20d8000b7fa9033fde29"} Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.453708 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4fe15275-4e76-46cc-9979-87af4f8a5407" (UID: "4fe15275-4e76-46cc-9979-87af4f8a5407"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.515894 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4fe15275-4e76-46cc-9979-87af4f8a5407" (UID: "4fe15275-4e76-46cc-9979-87af4f8a5407"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.530386 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.530427 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.682235 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4fe15275-4e76-46cc-9979-87af4f8a5407" (UID: "4fe15275-4e76-46cc-9979-87af4f8a5407"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.740411 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-config" (OuterVolumeSpecName: "config") pod "4fe15275-4e76-46cc-9979-87af4f8a5407" (UID: "4fe15275-4e76-46cc-9979-87af4f8a5407"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.745240 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.745276 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.777126 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4fe15275-4e76-46cc-9979-87af4f8a5407" (UID: "4fe15275-4e76-46cc-9979-87af4f8a5407"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:06 crc kubenswrapper[4677]: I1203 14:11:06.856178 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fe15275-4e76-46cc-9979-87af4f8a5407-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.012067 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57f57f74bc-885jf"] Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.033684 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57f57f74bc-885jf"] Dec 03 14:11:07 crc kubenswrapper[4677]: E1203 14:11:07.106517 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:11:07 crc kubenswrapper[4677]: E1203 14:11:07.107904 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:11:07 crc kubenswrapper[4677]: E1203 14:11:07.113303 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 14:11:07 crc kubenswrapper[4677]: E1203 14:11:07.113379 4677 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.406024 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cf99b568-bhfkp" event={"ID":"6c52f0ed-484d-44cf-85a3-17732f779171","Type":"ContainerStarted","Data":"bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996"} Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.407033 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.407072 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.414604 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d5f483c-fe0a-448f-be69-5c6bedf0e235","Type":"ContainerStarted","Data":"86fc75f6d9e78d73483665764065005aa6f9ae7a21f06cc8ae513c89c14d2af3"} Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.414760 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api-log" containerID="cri-o://cef027ba74898040c24a79f67c8c3f4c331596839de6a2032407493d873047b6" gracePeriod=30 Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.415022 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.415055 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api" containerID="cri-o://86fc75f6d9e78d73483665764065005aa6f9ae7a21f06cc8ae513c89c14d2af3" gracePeriod=30 Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.438823 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f","Type":"ContainerStarted","Data":"f6564e26530d6c61b84ca906f513da9db3069b88d42d7440279ab397a3af46bf"} Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.441312 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5cf99b568-bhfkp" podStartSLOduration=5.441294471 podStartE2EDuration="5.441294471s" podCreationTimestamp="2025-12-03 14:11:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:07.424958193 +0000 UTC m=+1458.171290648" watchObservedRunningTime="2025-12-03 14:11:07.441294471 +0000 UTC m=+1458.187626926" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.441853 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" event={"ID":"8e1c647c-16bc-4421-9bc8-7395212da3fe","Type":"ContainerStarted","Data":"b64ed340e3bc0b3fa5528dc8e8f644e44880f1b110f59e8974d6330b55aeff0a"} Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.442011 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.461493 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.4614702489999996 podStartE2EDuration="6.461470249s" podCreationTimestamp="2025-12-03 14:11:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:07.456335224 +0000 UTC m=+1458.202667699" watchObservedRunningTime="2025-12-03 14:11:07.461470249 +0000 UTC m=+1458.207802704" Dec 03 14:11:07 crc kubenswrapper[4677]: I1203 14:11:07.485375 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" podStartSLOduration=5.485354634 podStartE2EDuration="5.485354634s" podCreationTimestamp="2025-12-03 14:11:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:07.477782456 +0000 UTC m=+1458.224114911" watchObservedRunningTime="2025-12-03 14:11:07.485354634 +0000 UTC m=+1458.231687089" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.006034 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fe15275-4e76-46cc-9979-87af4f8a5407" path="/var/lib/kubelet/pods/4fe15275-4e76-46cc-9979-87af4f8a5407/volumes" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.054585 4677 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod991ccfb3-a839-473d-9c0b-874b5ff7fe60"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod991ccfb3-a839-473d-9c0b-874b5ff7fe60] : Timed out while waiting for systemd to remove kubepods-besteffort-pod991ccfb3_a839_473d_9c0b_874b5ff7fe60.slice" Dec 03 14:11:08 crc kubenswrapper[4677]: E1203 14:11:08.054643 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod991ccfb3-a839-473d-9c0b-874b5ff7fe60] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod991ccfb3-a839-473d-9c0b-874b5ff7fe60] : Timed out while waiting for systemd to remove kubepods-besteffort-pod991ccfb3_a839_473d_9c0b_874b5ff7fe60.slice" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" podUID="991ccfb3-a839-473d-9c0b-874b5ff7fe60" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.480808 4677 generic.go:334] "Generic (PLEG): container finished" podID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerID="86fc75f6d9e78d73483665764065005aa6f9ae7a21f06cc8ae513c89c14d2af3" exitCode=0 Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.480839 4677 generic.go:334] "Generic (PLEG): container finished" podID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerID="cef027ba74898040c24a79f67c8c3f4c331596839de6a2032407493d873047b6" exitCode=143 Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.480907 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d5f483c-fe0a-448f-be69-5c6bedf0e235","Type":"ContainerDied","Data":"86fc75f6d9e78d73483665764065005aa6f9ae7a21f06cc8ae513c89c14d2af3"} Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.480932 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d5f483c-fe0a-448f-be69-5c6bedf0e235","Type":"ContainerDied","Data":"cef027ba74898040c24a79f67c8c3f4c331596839de6a2032407493d873047b6"} Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.483110 4677 generic.go:334] "Generic (PLEG): container finished" podID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerID="2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0" exitCode=1 Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.483249 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.483738 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerDied","Data":"2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0"} Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.483790 4677 scope.go:117] "RemoveContainer" containerID="fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.483744 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.484565 4677 scope.go:117] "RemoveContainer" containerID="2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0" Dec 03 14:11:08 crc kubenswrapper[4677]: E1203 14:11:08.484749 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f93ff7e-f6ab-4c00-8284-70a4354e576a)\"" pod="openstack/watcher-decision-engine-0" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.576996 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c"] Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.584278 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f9c8fbfbf-rvh4c"] Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.771416 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774174 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-scripts\") pod \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774223 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d5f483c-fe0a-448f-be69-5c6bedf0e235-etc-machine-id\") pod \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774269 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data\") pod \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774288 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b47k9\" (UniqueName: \"kubernetes.io/projected/0d5f483c-fe0a-448f-be69-5c6bedf0e235-kube-api-access-b47k9\") pod \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774359 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data-custom\") pod \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774433 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-combined-ca-bundle\") pod \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774498 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d5f483c-fe0a-448f-be69-5c6bedf0e235-logs\") pod \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\" (UID: \"0d5f483c-fe0a-448f-be69-5c6bedf0e235\") " Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.774893 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d5f483c-fe0a-448f-be69-5c6bedf0e235-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0d5f483c-fe0a-448f-be69-5c6bedf0e235" (UID: "0d5f483c-fe0a-448f-be69-5c6bedf0e235"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.775083 4677 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d5f483c-fe0a-448f-be69-5c6bedf0e235-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.779628 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0d5f483c-fe0a-448f-be69-5c6bedf0e235" (UID: "0d5f483c-fe0a-448f-be69-5c6bedf0e235"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.781122 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d5f483c-fe0a-448f-be69-5c6bedf0e235-logs" (OuterVolumeSpecName: "logs") pod "0d5f483c-fe0a-448f-be69-5c6bedf0e235" (UID: "0d5f483c-fe0a-448f-be69-5c6bedf0e235"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.783556 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-scripts" (OuterVolumeSpecName: "scripts") pod "0d5f483c-fe0a-448f-be69-5c6bedf0e235" (UID: "0d5f483c-fe0a-448f-be69-5c6bedf0e235"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.801069 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d5f483c-fe0a-448f-be69-5c6bedf0e235-kube-api-access-b47k9" (OuterVolumeSpecName: "kube-api-access-b47k9") pod "0d5f483c-fe0a-448f-be69-5c6bedf0e235" (UID: "0d5f483c-fe0a-448f-be69-5c6bedf0e235"). InnerVolumeSpecName "kube-api-access-b47k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.839826 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d5f483c-fe0a-448f-be69-5c6bedf0e235" (UID: "0d5f483c-fe0a-448f-be69-5c6bedf0e235"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.880751 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d5f483c-fe0a-448f-be69-5c6bedf0e235-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.880792 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.880807 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b47k9\" (UniqueName: \"kubernetes.io/projected/0d5f483c-fe0a-448f-be69-5c6bedf0e235-kube-api-access-b47k9\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.880819 4677 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.880831 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.901111 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data" (OuterVolumeSpecName: "config-data") pod "0d5f483c-fe0a-448f-be69-5c6bedf0e235" (UID: "0d5f483c-fe0a-448f-be69-5c6bedf0e235"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:08 crc kubenswrapper[4677]: I1203 14:11:08.982285 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d5f483c-fe0a-448f-be69-5c6bedf0e235-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.126938 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6db8474b64-zrqnv"] Dec 03 14:11:09 crc kubenswrapper[4677]: E1203 14:11:09.127309 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api-log" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.127326 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api-log" Dec 03 14:11:09 crc kubenswrapper[4677]: E1203 14:11:09.127359 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.127366 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api" Dec 03 14:11:09 crc kubenswrapper[4677]: E1203 14:11:09.127377 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe15275-4e76-46cc-9979-87af4f8a5407" containerName="init" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.127384 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe15275-4e76-46cc-9979-87af4f8a5407" containerName="init" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.127559 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api-log" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.127569 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fe15275-4e76-46cc-9979-87af4f8a5407" containerName="init" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.127579 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" containerName="cinder-api" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.128501 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.135212 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.135374 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.160533 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.160674 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.171911 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6db8474b64-zrqnv"] Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.186111 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-config-data\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.186390 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkl7n\" (UniqueName: \"kubernetes.io/projected/9c952b2e-f0e2-40a8-b5ad-16965795c194-kube-api-access-zkl7n\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.186524 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-public-tls-certs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.186660 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-combined-ca-bundle\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.186789 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-config-data-custom\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.186898 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c952b2e-f0e2-40a8-b5ad-16965795c194-logs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.191516 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-internal-tls-certs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.292881 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-config-data-custom\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.292936 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c952b2e-f0e2-40a8-b5ad-16965795c194-logs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.293018 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-internal-tls-certs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.293051 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-config-data\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.293076 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkl7n\" (UniqueName: \"kubernetes.io/projected/9c952b2e-f0e2-40a8-b5ad-16965795c194-kube-api-access-zkl7n\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.293111 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-public-tls-certs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.293167 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-combined-ca-bundle\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.294148 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c952b2e-f0e2-40a8-b5ad-16965795c194-logs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.297670 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-config-data\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.301262 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-combined-ca-bundle\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.301579 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-internal-tls-certs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.302509 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-config-data-custom\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.311084 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c952b2e-f0e2-40a8-b5ad-16965795c194-public-tls-certs\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.315396 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkl7n\" (UniqueName: \"kubernetes.io/projected/9c952b2e-f0e2-40a8-b5ad-16965795c194-kube-api-access-zkl7n\") pod \"barbican-api-6db8474b64-zrqnv\" (UID: \"9c952b2e-f0e2-40a8-b5ad-16965795c194\") " pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.507916 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.537618 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.547147 4677 generic.go:334] "Generic (PLEG): container finished" podID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerID="5f8248e38c43d4427d7c586ea1b85645e3cbc9ad237b8dd3d5ef2c546003a0bc" exitCode=0 Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.547203 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerDied","Data":"5f8248e38c43d4427d7c586ea1b85645e3cbc9ad237b8dd3d5ef2c546003a0bc"} Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.549412 4677 scope.go:117] "RemoveContainer" containerID="2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0" Dec 03 14:11:09 crc kubenswrapper[4677]: E1203 14:11:09.549630 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f93ff7e-f6ab-4c00-8284-70a4354e576a)\"" pod="openstack/watcher-decision-engine-0" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.552905 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d5f483c-fe0a-448f-be69-5c6bedf0e235","Type":"ContainerDied","Data":"c8cea9e4f23cc2907255a9e1929a3c9249c9b5bd840231ea01c334295c477dc3"} Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.553110 4677 scope.go:117] "RemoveContainer" containerID="86fc75f6d9e78d73483665764065005aa6f9ae7a21f06cc8ae513c89c14d2af3" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.552968 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.559682 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" event={"ID":"9f174682-8786-4f0e-8ce3-f72c5560bde6","Type":"ContainerStarted","Data":"ca75f88a25f8cdb0d558af45508e079ebe05ef41bf4aa086532358d9e23a0b02"} Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.563262 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" event={"ID":"e5960f5e-97d6-45c7-b597-b7ae61478585","Type":"ContainerStarted","Data":"768ac4406683749207ca2d30b00d80a18bf20dcba0727896b63648cce1c73c4b"} Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.563290 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" event={"ID":"e5960f5e-97d6-45c7-b597-b7ae61478585","Type":"ContainerStarted","Data":"3b9b893f0f723dd03cbbfca7bbd179ce421bddf3de9776785946ee62c520a72b"} Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.598287 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-64fbb7dcd5-9hc8s" podStartSLOduration=3.183071921 podStartE2EDuration="7.598269751s" podCreationTimestamp="2025-12-03 14:11:02 +0000 UTC" firstStartedPulling="2025-12-03 14:11:04.306718976 +0000 UTC m=+1455.053051431" lastFinishedPulling="2025-12-03 14:11:08.721916806 +0000 UTC m=+1459.468249261" observedRunningTime="2025-12-03 14:11:09.596194856 +0000 UTC m=+1460.342527321" watchObservedRunningTime="2025-12-03 14:11:09.598269751 +0000 UTC m=+1460.344602216" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.642939 4677 scope.go:117] "RemoveContainer" containerID="cef027ba74898040c24a79f67c8c3f4c331596839de6a2032407493d873047b6" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.693737 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.710282 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.733074 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.735288 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.737253 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.744661 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.746061 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.771541 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.839543 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-55bcdc4f7c-89g9k" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.906194 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5574fd4bb8-fs46f"] Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.906415 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5574fd4bb8-fs46f" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-api" containerID="cri-o://2b1d3f58d578305185c5dc629c41dbf1e2307cf556f7b0097c814545841fbbfa" gracePeriod=30 Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.906803 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5574fd4bb8-fs46f" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-httpd" containerID="cri-o://4326918e7f551aea44b394c6bfbd72322962e712dde1ff6592bd906c64244143" gracePeriod=30 Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.917749 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.917788 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.917853 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-scripts\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.917873 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0312747f-a587-41d5-883b-748b2b42e1ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.918071 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.918437 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-config-data\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.918489 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.918515 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0312747f-a587-41d5-883b-748b2b42e1ea-logs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:09 crc kubenswrapper[4677]: I1203 14:11:09.918533 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c755f\" (UniqueName: \"kubernetes.io/projected/0312747f-a587-41d5-883b-748b2b42e1ea-kube-api-access-c755f\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.009344 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d5f483c-fe0a-448f-be69-5c6bedf0e235" path="/var/lib/kubelet/pods/0d5f483c-fe0a-448f-be69-5c6bedf0e235/volumes" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.013589 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="991ccfb3-a839-473d-9c0b-874b5ff7fe60" path="/var/lib/kubelet/pods/991ccfb3-a839-473d-9c0b-874b5ff7fe60/volumes" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021442 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021491 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021507 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-scripts\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021525 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0312747f-a587-41d5-883b-748b2b42e1ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021565 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021651 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-config-data\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021676 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021690 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0312747f-a587-41d5-883b-748b2b42e1ea-logs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.021710 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c755f\" (UniqueName: \"kubernetes.io/projected/0312747f-a587-41d5-883b-748b2b42e1ea-kube-api-access-c755f\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.027895 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-scripts\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.028417 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0312747f-a587-41d5-883b-748b2b42e1ea-logs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.028516 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0312747f-a587-41d5-883b-748b2b42e1ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.030000 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-config-data\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.032902 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.033484 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.051836 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.054516 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c755f\" (UniqueName: \"kubernetes.io/projected/0312747f-a587-41d5-883b-748b2b42e1ea-kube-api-access-c755f\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.065832 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0312747f-a587-41d5-883b-748b2b42e1ea-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0312747f-a587-41d5-883b-748b2b42e1ea\") " pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: W1203 14:11:10.100754 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f93ff7e_f6ab_4c00_8284_70a4354e576a.slice/crio-fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf.scope WatchSource:0}: Error finding container fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf: Status 404 returned error can't find the container with id fe18f0db60791f17798eb713d26fd01ec45968c9b8473ad807ca0e1fd33d2eaf Dec 03 14:11:10 crc kubenswrapper[4677]: W1203 14:11:10.162589 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fe15275_4e76_46cc_9979_87af4f8a5407.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fe15275_4e76_46cc_9979_87af4f8a5407.slice: no such file or directory Dec 03 14:11:10 crc kubenswrapper[4677]: W1203 14:11:10.162714 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d5f483c_fe0a_448f_be69_5c6bedf0e235.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d5f483c_fe0a_448f_be69_5c6bedf0e235.slice: no such file or directory Dec 03 14:11:10 crc kubenswrapper[4677]: W1203 14:11:10.211079 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f93ff7e_f6ab_4c00_8284_70a4354e576a.slice/crio-conmon-2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f93ff7e_f6ab_4c00_8284_70a4354e576a.slice/crio-conmon-2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0.scope: no such file or directory Dec 03 14:11:10 crc kubenswrapper[4677]: W1203 14:11:10.211133 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f93ff7e_f6ab_4c00_8284_70a4354e576a.slice/crio-2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f93ff7e_f6ab_4c00_8284_70a4354e576a.slice/crio-2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0.scope: no such file or directory Dec 03 14:11:10 crc kubenswrapper[4677]: W1203 14:11:10.232481 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f345427_301d_44b0_8e50_3d5e2444a48f.slice/crio-conmon-baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f345427_301d_44b0_8e50_3d5e2444a48f.slice/crio-conmon-baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1.scope: no such file or directory Dec 03 14:11:10 crc kubenswrapper[4677]: W1203 14:11:10.232521 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f345427_301d_44b0_8e50_3d5e2444a48f.slice/crio-baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f345427_301d_44b0_8e50_3d5e2444a48f.slice/crio-baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1.scope: no such file or directory Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.270504 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.308113 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6db8474b64-zrqnv"] Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.363321 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.448497 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-sg-core-conf-yaml\") pod \"2f345427-301d-44b0-8e50-3d5e2444a48f\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.448571 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-combined-ca-bundle\") pod \"2f345427-301d-44b0-8e50-3d5e2444a48f\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.448626 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-config-data\") pod \"2f345427-301d-44b0-8e50-3d5e2444a48f\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.448763 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-scripts\") pod \"2f345427-301d-44b0-8e50-3d5e2444a48f\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.448841 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-run-httpd\") pod \"2f345427-301d-44b0-8e50-3d5e2444a48f\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.448869 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hglx2\" (UniqueName: \"kubernetes.io/projected/2f345427-301d-44b0-8e50-3d5e2444a48f-kube-api-access-hglx2\") pod \"2f345427-301d-44b0-8e50-3d5e2444a48f\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.449040 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-log-httpd\") pod \"2f345427-301d-44b0-8e50-3d5e2444a48f\" (UID: \"2f345427-301d-44b0-8e50-3d5e2444a48f\") " Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.450132 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f345427-301d-44b0-8e50-3d5e2444a48f" (UID: "2f345427-301d-44b0-8e50-3d5e2444a48f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.454277 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f345427-301d-44b0-8e50-3d5e2444a48f" (UID: "2f345427-301d-44b0-8e50-3d5e2444a48f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.462358 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-scripts" (OuterVolumeSpecName: "scripts") pod "2f345427-301d-44b0-8e50-3d5e2444a48f" (UID: "2f345427-301d-44b0-8e50-3d5e2444a48f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.465828 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f345427-301d-44b0-8e50-3d5e2444a48f-kube-api-access-hglx2" (OuterVolumeSpecName: "kube-api-access-hglx2") pod "2f345427-301d-44b0-8e50-3d5e2444a48f" (UID: "2f345427-301d-44b0-8e50-3d5e2444a48f"). InnerVolumeSpecName "kube-api-access-hglx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.533248 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f345427-301d-44b0-8e50-3d5e2444a48f" (UID: "2f345427-301d-44b0-8e50-3d5e2444a48f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:10 crc kubenswrapper[4677]: E1203 14:11:10.542931 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25625b10_c45e_40f0_b342_bdcdc3e79c9c.slice/crio-0e8086cca4d17a62b570dc85b637b50f072e46f99feb791e5c32cd1c611f2cfc\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e5be0cc_a23f_4d36_a5a1_405d14814402.slice/crio-conmon-4326918e7f551aea44b394c6bfbd72322962e712dde1ff6592bd906c64244143.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5edce1f1_9c11_4049_b286_a16939acb2e7.slice/crio-conmon-e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5edce1f1_9c11_4049_b286_a16939acb2e7.slice/crio-e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fd09b0c_1690_42e1_b97a_a18e7fefc813.slice/crio-64425cc924e22eaa6e7d652322613611f220ab3d76f8ce2539e1804f9ab5321a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f345427_301d_44b0_8e50_3d5e2444a48f.slice/crio-1ec3750646ee7fe2cd08896eb5a41aea1e9aae3c83da20d8000b7fa9033fde29.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ec3794_7d5b_49dc_a957_bee4761d55f0.slice/crio-8a66e68be6fff49b84696df5494a916693c024f8458f6735679bc7c8ba30309a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fd09b0c_1690_42e1_b97a_a18e7fefc813.slice/crio-f99039f40832f2e19ff5f6dfbade0ff5509753b5d47c31912f7113c43996b123\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fd09b0c_1690_42e1_b97a_a18e7fefc813.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ec3794_7d5b_49dc_a957_bee4761d55f0.slice/crio-conmon-8a66e68be6fff49b84696df5494a916693c024f8458f6735679bc7c8ba30309a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f345427_301d_44b0_8e50_3d5e2444a48f.slice/crio-conmon-5f8248e38c43d4427d7c586ea1b85645e3cbc9ad237b8dd3d5ef2c546003a0bc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f345427_301d_44b0_8e50_3d5e2444a48f.slice/crio-conmon-1ec3750646ee7fe2cd08896eb5a41aea1e9aae3c83da20d8000b7fa9033fde29.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25625b10_c45e_40f0_b342_bdcdc3e79c9c.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.551090 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.551121 4677 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.551133 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hglx2\" (UniqueName: \"kubernetes.io/projected/2f345427-301d-44b0-8e50-3d5e2444a48f-kube-api-access-hglx2\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.551146 4677 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f345427-301d-44b0-8e50-3d5e2444a48f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.551156 4677 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.604179 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db8474b64-zrqnv" event={"ID":"9c952b2e-f0e2-40a8-b5ad-16965795c194","Type":"ContainerStarted","Data":"e9dca2ed503453096ed31da876987d847d0e0a910ed4c8f05df99d31f54b51af"} Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.689905 4677 generic.go:334] "Generic (PLEG): container finished" podID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" exitCode=137 Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.689994 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5edce1f1-9c11-4049-b286-a16939acb2e7","Type":"ContainerDied","Data":"e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37"} Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.691088 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-config-data" (OuterVolumeSpecName: "config-data") pod "2f345427-301d-44b0-8e50-3d5e2444a48f" (UID: "2f345427-301d-44b0-8e50-3d5e2444a48f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.815507 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.860801 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f","Type":"ContainerStarted","Data":"8886485c653c22a40e3e8fc1b00e4bf5fc99929a49f0b665558ae0b9efda43ee"} Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.885495 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f345427-301d-44b0-8e50-3d5e2444a48f" (UID: "2f345427-301d-44b0-8e50-3d5e2444a48f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.894201 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" event={"ID":"9f174682-8786-4f0e-8ce3-f72c5560bde6","Type":"ContainerStarted","Data":"8176d0ed4224f232def3205e30e8e79d8df882c8d61029953acbbac01d3a5fe3"} Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.918586 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f345427-301d-44b0-8e50-3d5e2444a48f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.951229 4677 generic.go:334] "Generic (PLEG): container finished" podID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerID="4326918e7f551aea44b394c6bfbd72322962e712dde1ff6592bd906c64244143" exitCode=0 Dec 03 14:11:10 crc kubenswrapper[4677]: I1203 14:11:10.951381 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5574fd4bb8-fs46f" event={"ID":"1e5be0cc-a23f-4d36-a5a1-405d14814402","Type":"ContainerDied","Data":"4326918e7f551aea44b394c6bfbd72322962e712dde1ff6592bd906c64244143"} Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.011923 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.014121 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f345427-301d-44b0-8e50-3d5e2444a48f","Type":"ContainerDied","Data":"9e7f7fe92ce681f771cb76ea336c3d9d35096f5c7f447611daa512ad35854582"} Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.014180 4677 scope.go:117] "RemoveContainer" containerID="baef2707fd36811f2535afb78dd8e0977c51109417f24983cacc84029e3c14e1" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.054277 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=8.512843197 podStartE2EDuration="10.054256306s" podCreationTimestamp="2025-12-03 14:11:01 +0000 UTC" firstStartedPulling="2025-12-03 14:11:03.502753646 +0000 UTC m=+1454.249086101" lastFinishedPulling="2025-12-03 14:11:05.044166755 +0000 UTC m=+1455.790499210" observedRunningTime="2025-12-03 14:11:10.935693102 +0000 UTC m=+1461.682025587" watchObservedRunningTime="2025-12-03 14:11:11.054256306 +0000 UTC m=+1461.800588751" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.085207 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-67bc86b468-qwl7k" podStartSLOduration=4.584293482 podStartE2EDuration="9.085185295s" podCreationTimestamp="2025-12-03 14:11:02 +0000 UTC" firstStartedPulling="2025-12-03 14:11:04.216342871 +0000 UTC m=+1454.962675326" lastFinishedPulling="2025-12-03 14:11:08.717234684 +0000 UTC m=+1459.463567139" observedRunningTime="2025-12-03 14:11:10.973735178 +0000 UTC m=+1461.720067623" watchObservedRunningTime="2025-12-03 14:11:11.085185295 +0000 UTC m=+1461.831517750" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.145306 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.169582 4677 scope.go:117] "RemoveContainer" containerID="1ec3750646ee7fe2cd08896eb5a41aea1e9aae3c83da20d8000b7fa9033fde29" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.201672 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5edce1f1-9c11-4049-b286-a16939acb2e7-logs\") pod \"5edce1f1-9c11-4049-b286-a16939acb2e7\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.202024 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-combined-ca-bundle\") pod \"5edce1f1-9c11-4049-b286-a16939acb2e7\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.202267 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-config-data\") pod \"5edce1f1-9c11-4049-b286-a16939acb2e7\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.202288 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvqh6\" (UniqueName: \"kubernetes.io/projected/5edce1f1-9c11-4049-b286-a16939acb2e7-kube-api-access-jvqh6\") pod \"5edce1f1-9c11-4049-b286-a16939acb2e7\" (UID: \"5edce1f1-9c11-4049-b286-a16939acb2e7\") " Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.205915 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5edce1f1-9c11-4049-b286-a16939acb2e7-logs" (OuterVolumeSpecName: "logs") pod "5edce1f1-9c11-4049-b286-a16939acb2e7" (UID: "5edce1f1-9c11-4049-b286-a16939acb2e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.211302 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5edce1f1-9c11-4049-b286-a16939acb2e7-kube-api-access-jvqh6" (OuterVolumeSpecName: "kube-api-access-jvqh6") pod "5edce1f1-9c11-4049-b286-a16939acb2e7" (UID: "5edce1f1-9c11-4049-b286-a16939acb2e7"). InnerVolumeSpecName "kube-api-access-jvqh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.225203 4677 scope.go:117] "RemoveContainer" containerID="5f8248e38c43d4427d7c586ea1b85645e3cbc9ad237b8dd3d5ef2c546003a0bc" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.280339 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5edce1f1-9c11-4049-b286-a16939acb2e7" (UID: "5edce1f1-9c11-4049-b286-a16939acb2e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.305617 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvqh6\" (UniqueName: \"kubernetes.io/projected/5edce1f1-9c11-4049-b286-a16939acb2e7-kube-api-access-jvqh6\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.305654 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5edce1f1-9c11-4049-b286-a16939acb2e7-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.305663 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.323224 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.402452 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.406990 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-config-data" (OuterVolumeSpecName: "config-data") pod "5edce1f1-9c11-4049-b286-a16939acb2e7" (UID: "5edce1f1-9c11-4049-b286-a16939acb2e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.408403 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5edce1f1-9c11-4049-b286-a16939acb2e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.434663 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.443237 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:11 crc kubenswrapper[4677]: E1203 14:11:11.443721 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.443742 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:11:11 crc kubenswrapper[4677]: E1203 14:11:11.443776 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="sg-core" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.443782 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="sg-core" Dec 03 14:11:11 crc kubenswrapper[4677]: E1203 14:11:11.443796 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="proxy-httpd" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.443801 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="proxy-httpd" Dec 03 14:11:11 crc kubenswrapper[4677]: E1203 14:11:11.443811 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="ceilometer-notification-agent" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.443816 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="ceilometer-notification-agent" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.444030 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="ceilometer-notification-agent" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.444045 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="proxy-httpd" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.444056 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" containerName="watcher-applier" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.444070 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" containerName="sg-core" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.449115 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.451369 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.452104 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.461033 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.468488 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.511155 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-log-httpd\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.511221 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-config-data\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.511301 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-run-httpd\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.511417 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxmw6\" (UniqueName: \"kubernetes.io/projected/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-kube-api-access-xxmw6\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.511442 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.511502 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.511535 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-scripts\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.550595 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.560879 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.173:8080/\": dial tcp 10.217.0.173:8080: connect: connection refused" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.613706 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-run-httpd\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.613843 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxmw6\" (UniqueName: \"kubernetes.io/projected/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-kube-api-access-xxmw6\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.613872 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.613921 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.613996 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-scripts\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.614063 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-log-httpd\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.614086 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-config-data\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.617374 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-run-httpd\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.617903 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-log-httpd\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.619303 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.622111 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-config-data\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.622242 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-scripts\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.624494 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.635883 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxmw6\" (UniqueName: \"kubernetes.io/projected/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-kube-api-access-xxmw6\") pod \"ceilometer-0\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.788187 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:11 crc kubenswrapper[4677]: I1203 14:11:11.993976 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f345427-301d-44b0-8e50-3d5e2444a48f" path="/var/lib/kubelet/pods/2f345427-301d-44b0-8e50-3d5e2444a48f/volumes" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.051473 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5edce1f1-9c11-4049-b286-a16939acb2e7","Type":"ContainerDied","Data":"23b01ff80ae7cb01f2e27c8da418d73a2ade4176ea17e83f6742857024b82166"} Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.051533 4677 scope.go:117] "RemoveContainer" containerID="e0bc2742cdf9016c792dd9113bbd48b907e450e53daa18f9f2c8345f2937ac37" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.051672 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.067989 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0312747f-a587-41d5-883b-748b2b42e1ea","Type":"ContainerStarted","Data":"f0fd9f3a906cc93806dc718e280cc0a67a0e3b637d5c929b05677484eddbd3e2"} Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.077682 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.096580 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.097362 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db8474b64-zrqnv" event={"ID":"9c952b2e-f0e2-40a8-b5ad-16965795c194","Type":"ContainerStarted","Data":"70162db1c07185bd20a1c3828aaf2a59a24cb52349f029283bc7f1e78c11f523"} Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.097391 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db8474b64-zrqnv" event={"ID":"9c952b2e-f0e2-40a8-b5ad-16965795c194","Type":"ContainerStarted","Data":"95831e25abf07165423c63cf21f165d7b39b4cf740185ef113806a4eb7bddca7"} Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.115577 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.123086 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.129764 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.130285 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.136343 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6db8474b64-zrqnv" podStartSLOduration=3.136321334 podStartE2EDuration="3.136321334s" podCreationTimestamp="2025-12-03 14:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:12.125420838 +0000 UTC m=+1462.871753303" watchObservedRunningTime="2025-12-03 14:11:12.136321334 +0000 UTC m=+1462.882653789" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.228128 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlj8k\" (UniqueName: \"kubernetes.io/projected/2c0001c6-665a-4888-a2ff-02819272545c-kube-api-access-vlj8k\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.228186 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0001c6-665a-4888-a2ff-02819272545c-logs\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.228328 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0001c6-665a-4888-a2ff-02819272545c-config-data\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.228367 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0001c6-665a-4888-a2ff-02819272545c-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.303701 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:12 crc kubenswrapper[4677]: W1203 14:11:12.313502 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod257c64e3_7a6b_4d9f_8a08_6aaea8dc9f5b.slice/crio-7f57059feeda29c95bd51bb86eb93d2120e8911f2657d2673e90ba647db0afdf WatchSource:0}: Error finding container 7f57059feeda29c95bd51bb86eb93d2120e8911f2657d2673e90ba647db0afdf: Status 404 returned error can't find the container with id 7f57059feeda29c95bd51bb86eb93d2120e8911f2657d2673e90ba647db0afdf Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.329915 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlj8k\" (UniqueName: \"kubernetes.io/projected/2c0001c6-665a-4888-a2ff-02819272545c-kube-api-access-vlj8k\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.330226 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0001c6-665a-4888-a2ff-02819272545c-logs\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.330399 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0001c6-665a-4888-a2ff-02819272545c-config-data\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.330435 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0001c6-665a-4888-a2ff-02819272545c-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.331069 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c0001c6-665a-4888-a2ff-02819272545c-logs\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.339308 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c0001c6-665a-4888-a2ff-02819272545c-config-data\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.342469 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0001c6-665a-4888-a2ff-02819272545c-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.351252 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlj8k\" (UniqueName: \"kubernetes.io/projected/2c0001c6-665a-4888-a2ff-02819272545c-kube-api-access-vlj8k\") pod \"watcher-applier-0\" (UID: \"2c0001c6-665a-4888-a2ff-02819272545c\") " pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.454657 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.894135 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.976997 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7857f7fffc-qb4s8"] Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.977257 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" podUID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerName="dnsmasq-dns" containerID="cri-o://f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36" gracePeriod=10 Dec 03 14:11:12 crc kubenswrapper[4677]: I1203 14:11:12.988173 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 14:11:12 crc kubenswrapper[4677]: W1203 14:11:12.998191 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c0001c6_665a_4888_a2ff_02819272545c.slice/crio-7c8ab23d916200d23c7ec8a0d50e2efc6d90d3eeed968d5f16f312ff6c09b3ee WatchSource:0}: Error finding container 7c8ab23d916200d23c7ec8a0d50e2efc6d90d3eeed968d5f16f312ff6c09b3ee: Status 404 returned error can't find the container with id 7c8ab23d916200d23c7ec8a0d50e2efc6d90d3eeed968d5f16f312ff6c09b3ee Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.173718 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"2c0001c6-665a-4888-a2ff-02819272545c","Type":"ContainerStarted","Data":"7c8ab23d916200d23c7ec8a0d50e2efc6d90d3eeed968d5f16f312ff6c09b3ee"} Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.205174 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerStarted","Data":"7f57059feeda29c95bd51bb86eb93d2120e8911f2657d2673e90ba647db0afdf"} Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.241679 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0312747f-a587-41d5-883b-748b2b42e1ea","Type":"ContainerStarted","Data":"6083d9d1c4c2737ec5a86ef3a6454be17c2498624ee94c816a7f7943da39913a"} Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.241790 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.241855 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.601213 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.669510 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-config\") pod \"fa7ee2cb-a257-4314-b745-788aa64ca1de\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.669570 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkr8v\" (UniqueName: \"kubernetes.io/projected/fa7ee2cb-a257-4314-b745-788aa64ca1de-kube-api-access-zkr8v\") pod \"fa7ee2cb-a257-4314-b745-788aa64ca1de\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.669617 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-sb\") pod \"fa7ee2cb-a257-4314-b745-788aa64ca1de\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.669642 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-nb\") pod \"fa7ee2cb-a257-4314-b745-788aa64ca1de\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.669802 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-swift-storage-0\") pod \"fa7ee2cb-a257-4314-b745-788aa64ca1de\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.669825 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-svc\") pod \"fa7ee2cb-a257-4314-b745-788aa64ca1de\" (UID: \"fa7ee2cb-a257-4314-b745-788aa64ca1de\") " Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.708437 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa7ee2cb-a257-4314-b745-788aa64ca1de-kube-api-access-zkr8v" (OuterVolumeSpecName: "kube-api-access-zkr8v") pod "fa7ee2cb-a257-4314-b745-788aa64ca1de" (UID: "fa7ee2cb-a257-4314-b745-788aa64ca1de"). InnerVolumeSpecName "kube-api-access-zkr8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.771453 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkr8v\" (UniqueName: \"kubernetes.io/projected/fa7ee2cb-a257-4314-b745-788aa64ca1de-kube-api-access-zkr8v\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.812658 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fa7ee2cb-a257-4314-b745-788aa64ca1de" (UID: "fa7ee2cb-a257-4314-b745-788aa64ca1de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.818616 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fa7ee2cb-a257-4314-b745-788aa64ca1de" (UID: "fa7ee2cb-a257-4314-b745-788aa64ca1de"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.831900 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fa7ee2cb-a257-4314-b745-788aa64ca1de" (UID: "fa7ee2cb-a257-4314-b745-788aa64ca1de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.836435 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fa7ee2cb-a257-4314-b745-788aa64ca1de" (UID: "fa7ee2cb-a257-4314-b745-788aa64ca1de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.859289 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-config" (OuterVolumeSpecName: "config") pod "fa7ee2cb-a257-4314-b745-788aa64ca1de" (UID: "fa7ee2cb-a257-4314-b745-788aa64ca1de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.872960 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.872987 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.872996 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.873006 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:13 crc kubenswrapper[4677]: I1203 14:11:13.873015 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa7ee2cb-a257-4314-b745-788aa64ca1de-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.005918 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5edce1f1-9c11-4049-b286-a16939acb2e7" path="/var/lib/kubelet/pods/5edce1f1-9c11-4049-b286-a16939acb2e7/volumes" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.286999 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"2c0001c6-665a-4888-a2ff-02819272545c","Type":"ContainerStarted","Data":"08d74d408c1e5fdf904ed4336462891163d15e6d80130c1b892d69ef7391cf5e"} Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.293609 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerStarted","Data":"5ad85e76961d44f7fe0734108239a0dc760a1819635d79bfaa56ce984ee7c150"} Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.293647 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerStarted","Data":"51481a69112e94804b51807a499203d9b51eadf16fad172012edf688f788270a"} Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.315280 4677 generic.go:334] "Generic (PLEG): container finished" podID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerID="f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36" exitCode=0 Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.315372 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" event={"ID":"fa7ee2cb-a257-4314-b745-788aa64ca1de","Type":"ContainerDied","Data":"f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36"} Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.315399 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" event={"ID":"fa7ee2cb-a257-4314-b745-788aa64ca1de","Type":"ContainerDied","Data":"f3e5abfef88e94510724c8ffa1fdff5fc5b011a0f51bca91069e2b058e7be20a"} Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.315416 4677 scope.go:117] "RemoveContainer" containerID="f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.315547 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7857f7fffc-qb4s8" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.339766 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=2.33974924 podStartE2EDuration="2.33974924s" podCreationTimestamp="2025-12-03 14:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:14.326149404 +0000 UTC m=+1465.072481869" watchObservedRunningTime="2025-12-03 14:11:14.33974924 +0000 UTC m=+1465.086081695" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.340348 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0312747f-a587-41d5-883b-748b2b42e1ea","Type":"ContainerStarted","Data":"8cd67175784decae4addfb46eba6598171bc6179c2b5663f204c266e851758b0"} Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.340890 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.362655 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7857f7fffc-qb4s8"] Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.362743 4677 scope.go:117] "RemoveContainer" containerID="1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.377698 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7857f7fffc-qb4s8"] Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.400276 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.400258373 podStartE2EDuration="5.400258373s" podCreationTimestamp="2025-12-03 14:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:14.391871184 +0000 UTC m=+1465.138203639" watchObservedRunningTime="2025-12-03 14:11:14.400258373 +0000 UTC m=+1465.146590828" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.408129 4677 scope.go:117] "RemoveContainer" containerID="f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36" Dec 03 14:11:14 crc kubenswrapper[4677]: E1203 14:11:14.409329 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36\": container with ID starting with f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36 not found: ID does not exist" containerID="f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.409375 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36"} err="failed to get container status \"f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36\": rpc error: code = NotFound desc = could not find container \"f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36\": container with ID starting with f69678275b9543c01d4e8050fc53fab98a618a149235d2c358257eddecf75c36 not found: ID does not exist" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.409402 4677 scope.go:117] "RemoveContainer" containerID="1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d" Dec 03 14:11:14 crc kubenswrapper[4677]: E1203 14:11:14.409655 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d\": container with ID starting with 1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d not found: ID does not exist" containerID="1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.409688 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d"} err="failed to get container status \"1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d\": rpc error: code = NotFound desc = could not find container \"1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d\": container with ID starting with 1ac176f47a8b385ef7b5cf6b3fa37dcb2be4c1d7de8d9b09758a2705687e7f8d not found: ID does not exist" Dec 03 14:11:14 crc kubenswrapper[4677]: I1203 14:11:14.897460 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.006849 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.349014 4677 generic.go:334] "Generic (PLEG): container finished" podID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerID="2b1d3f58d578305185c5dc629c41dbf1e2307cf556f7b0097c814545841fbbfa" exitCode=0 Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.349086 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5574fd4bb8-fs46f" event={"ID":"1e5be0cc-a23f-4d36-a5a1-405d14814402","Type":"ContainerDied","Data":"2b1d3f58d578305185c5dc629c41dbf1e2307cf556f7b0097c814545841fbbfa"} Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.350647 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerStarted","Data":"83aa0a8551635104eb88c6bbb4ba0d82ce1418736f16770c2b59eb12f5996807"} Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.722494 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.816382 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-ovndb-tls-certs\") pod \"1e5be0cc-a23f-4d36-a5a1-405d14814402\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.816440 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-combined-ca-bundle\") pod \"1e5be0cc-a23f-4d36-a5a1-405d14814402\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.816475 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-httpd-config\") pod \"1e5be0cc-a23f-4d36-a5a1-405d14814402\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.816546 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-config\") pod \"1e5be0cc-a23f-4d36-a5a1-405d14814402\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.816575 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb6j4\" (UniqueName: \"kubernetes.io/projected/1e5be0cc-a23f-4d36-a5a1-405d14814402-kube-api-access-nb6j4\") pod \"1e5be0cc-a23f-4d36-a5a1-405d14814402\" (UID: \"1e5be0cc-a23f-4d36-a5a1-405d14814402\") " Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.822297 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "1e5be0cc-a23f-4d36-a5a1-405d14814402" (UID: "1e5be0cc-a23f-4d36-a5a1-405d14814402"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.822922 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e5be0cc-a23f-4d36-a5a1-405d14814402-kube-api-access-nb6j4" (OuterVolumeSpecName: "kube-api-access-nb6j4") pod "1e5be0cc-a23f-4d36-a5a1-405d14814402" (UID: "1e5be0cc-a23f-4d36-a5a1-405d14814402"). InnerVolumeSpecName "kube-api-access-nb6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.871083 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85b5c78d96-8262j" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.883671 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e5be0cc-a23f-4d36-a5a1-405d14814402" (UID: "1e5be0cc-a23f-4d36-a5a1-405d14814402"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.910790 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "1e5be0cc-a23f-4d36-a5a1-405d14814402" (UID: "1e5be0cc-a23f-4d36-a5a1-405d14814402"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.918143 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-config" (OuterVolumeSpecName: "config") pod "1e5be0cc-a23f-4d36-a5a1-405d14814402" (UID: "1e5be0cc-a23f-4d36-a5a1-405d14814402"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.919091 4677 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.919188 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.919340 4677 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.919456 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5be0cc-a23f-4d36-a5a1-405d14814402-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.919555 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb6j4\" (UniqueName: \"kubernetes.io/projected/1e5be0cc-a23f-4d36-a5a1-405d14814402-kube-api-access-nb6j4\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:15 crc kubenswrapper[4677]: I1203 14:11:15.991305 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa7ee2cb-a257-4314-b745-788aa64ca1de" path="/var/lib/kubelet/pods/fa7ee2cb-a257-4314-b745-788aa64ca1de/volumes" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.316278 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.330061 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.371783 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5574fd4bb8-fs46f" event={"ID":"1e5be0cc-a23f-4d36-a5a1-405d14814402","Type":"ContainerDied","Data":"710ecce1ef0310635d23db5287532bc42741253c26e57418757aac09eb1e1647"} Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.371866 4677 scope.go:117] "RemoveContainer" containerID="4326918e7f551aea44b394c6bfbd72322962e712dde1ff6592bd906c64244143" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.371877 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5574fd4bb8-fs46f" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.374571 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerStarted","Data":"168f927552bf3aafd7201ec78f561e04ba823cb86d59c65a4a423d5e309d0451"} Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.399305 4677 scope.go:117] "RemoveContainer" containerID="2b1d3f58d578305185c5dc629c41dbf1e2307cf556f7b0097c814545841fbbfa" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.412059 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.421375 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5574fd4bb8-fs46f"] Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.439147 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5574fd4bb8-fs46f"] Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.450536 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.832859264 podStartE2EDuration="5.45050944s" podCreationTimestamp="2025-12-03 14:11:11 +0000 UTC" firstStartedPulling="2025-12-03 14:11:12.31764811 +0000 UTC m=+1463.063980565" lastFinishedPulling="2025-12-03 14:11:15.935298286 +0000 UTC m=+1466.681630741" observedRunningTime="2025-12-03 14:11:16.411665273 +0000 UTC m=+1467.157997728" watchObservedRunningTime="2025-12-03 14:11:16.45050944 +0000 UTC m=+1467.196841895" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.720884 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 14:11:16 crc kubenswrapper[4677]: I1203 14:11:16.778903 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:17 crc kubenswrapper[4677]: I1203 14:11:17.386529 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:11:17 crc kubenswrapper[4677]: I1203 14:11:17.386784 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="cinder-scheduler" containerID="cri-o://f6564e26530d6c61b84ca906f513da9db3069b88d42d7440279ab397a3af46bf" gracePeriod=30 Dec 03 14:11:17 crc kubenswrapper[4677]: I1203 14:11:17.386999 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="probe" containerID="cri-o://8886485c653c22a40e3e8fc1b00e4bf5fc99929a49f0b665558ae0b9efda43ee" gracePeriod=30 Dec 03 14:11:17 crc kubenswrapper[4677]: I1203 14:11:17.455037 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Dec 03 14:11:17 crc kubenswrapper[4677]: I1203 14:11:17.988630 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" path="/var/lib/kubelet/pods/1e5be0cc-a23f-4d36-a5a1-405d14814402/volumes" Dec 03 14:11:18 crc kubenswrapper[4677]: I1203 14:11:18.395434 4677 generic.go:334] "Generic (PLEG): container finished" podID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerID="8886485c653c22a40e3e8fc1b00e4bf5fc99929a49f0b665558ae0b9efda43ee" exitCode=0 Dec 03 14:11:18 crc kubenswrapper[4677]: I1203 14:11:18.396773 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f","Type":"ContainerDied","Data":"8886485c653c22a40e3e8fc1b00e4bf5fc99929a49f0b665558ae0b9efda43ee"} Dec 03 14:11:19 crc kubenswrapper[4677]: I1203 14:11:19.160380 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:19 crc kubenswrapper[4677]: I1203 14:11:19.160467 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:19 crc kubenswrapper[4677]: I1203 14:11:19.161130 4677 scope.go:117] "RemoveContainer" containerID="2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0" Dec 03 14:11:19 crc kubenswrapper[4677]: I1203 14:11:19.419588 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerStarted","Data":"bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77"} Dec 03 14:11:20 crc kubenswrapper[4677]: I1203 14:11:20.443462 4677 generic.go:334] "Generic (PLEG): container finished" podID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerID="f6564e26530d6c61b84ca906f513da9db3069b88d42d7440279ab397a3af46bf" exitCode=0 Dec 03 14:11:20 crc kubenswrapper[4677]: I1203 14:11:20.444497 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f","Type":"ContainerDied","Data":"f6564e26530d6c61b84ca906f513da9db3069b88d42d7440279ab397a3af46bf"} Dec 03 14:11:20 crc kubenswrapper[4677]: I1203 14:11:20.616387 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:11:20 crc kubenswrapper[4677]: I1203 14:11:20.789519 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67bf5cb8b6-4lwks" Dec 03 14:11:20 crc kubenswrapper[4677]: I1203 14:11:20.960369 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.126046 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-scripts\") pod \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.126401 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data-custom\") pod \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.126497 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data\") pod \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.126534 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5dsl\" (UniqueName: \"kubernetes.io/projected/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-kube-api-access-t5dsl\") pod \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.126556 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-combined-ca-bundle\") pod \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.126668 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-etc-machine-id\") pod \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\" (UID: \"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f\") " Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.127557 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" (UID: "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.129544 4677 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.141839 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-scripts" (OuterVolumeSpecName: "scripts") pod "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" (UID: "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.166816 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" (UID: "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.169145 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-kube-api-access-t5dsl" (OuterVolumeSpecName: "kube-api-access-t5dsl") pod "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" (UID: "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f"). InnerVolumeSpecName "kube-api-access-t5dsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.212255 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" (UID: "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.231547 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.231815 4677 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.231909 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5dsl\" (UniqueName: \"kubernetes.io/projected/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-kube-api-access-t5dsl\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.232024 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.308983 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data" (OuterVolumeSpecName: "config-data") pod "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" (UID: "46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.333500 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.454712 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.458042 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f","Type":"ContainerDied","Data":"a9762b43b134307683fdf5fe77343fee6a501c27cc95e4b82bc3d93cdad77824"} Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.458114 4677 scope.go:117] "RemoveContainer" containerID="8886485c653c22a40e3e8fc1b00e4bf5fc99929a49f0b665558ae0b9efda43ee" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.490480 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.495177 4677 scope.go:117] "RemoveContainer" containerID="f6564e26530d6c61b84ca906f513da9db3069b88d42d7440279ab397a3af46bf" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.504066 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.506172 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537004 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:21 crc kubenswrapper[4677]: E1203 14:11:21.537423 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="cinder-scheduler" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537441 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="cinder-scheduler" Dec 03 14:11:21 crc kubenswrapper[4677]: E1203 14:11:21.537455 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-httpd" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537462 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-httpd" Dec 03 14:11:21 crc kubenswrapper[4677]: E1203 14:11:21.537480 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-api" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537487 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-api" Dec 03 14:11:21 crc kubenswrapper[4677]: E1203 14:11:21.537504 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerName="init" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537511 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerName="init" Dec 03 14:11:21 crc kubenswrapper[4677]: E1203 14:11:21.537529 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="probe" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537536 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="probe" Dec 03 14:11:21 crc kubenswrapper[4677]: E1203 14:11:21.537547 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerName="dnsmasq-dns" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537552 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerName="dnsmasq-dns" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537718 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-httpd" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537735 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="cinder-scheduler" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537747 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa7ee2cb-a257-4314-b745-788aa64ca1de" containerName="dnsmasq-dns" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537755 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e5be0cc-a23f-4d36-a5a1-405d14814402" containerName="neutron-api" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.537765 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" containerName="probe" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.538760 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.542414 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.558845 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.603066 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6db8474b64-zrqnv" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.648787 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.649028 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.649203 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.649270 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-config-data\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.649454 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-scripts\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.649659 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc4lx\" (UniqueName: \"kubernetes.io/projected/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-kube-api-access-wc4lx\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.726929 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5cf99b568-bhfkp"] Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.728765 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5cf99b568-bhfkp" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api-log" containerID="cri-o://738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525" gracePeriod=30 Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.728923 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5cf99b568-bhfkp" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api" containerID="cri-o://bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996" gracePeriod=30 Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.754815 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc4lx\" (UniqueName: \"kubernetes.io/projected/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-kube-api-access-wc4lx\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.757078 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.757193 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.757291 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.757352 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-config-data\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.757446 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-scripts\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.761424 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.783696 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-scripts\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.784464 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.788473 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-config-data\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.790263 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc4lx\" (UniqueName: \"kubernetes.io/projected/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-kube-api-access-wc4lx\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.802629 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f5a100-06b3-4f3c-beb1-7df2d48b6bc0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0\") " pod="openstack/cinder-scheduler-0" Dec 03 14:11:21 crc kubenswrapper[4677]: I1203 14:11:21.868152 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:11:22 crc kubenswrapper[4677]: I1203 14:11:22.013814 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f" path="/var/lib/kubelet/pods/46aa8f58-3b08-4ddf-9e54-ef4b6098ba9f/volumes" Dec 03 14:11:22 crc kubenswrapper[4677]: I1203 14:11:22.458256 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Dec 03 14:11:22 crc kubenswrapper[4677]: I1203 14:11:22.510446 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Dec 03 14:11:22 crc kubenswrapper[4677]: I1203 14:11:22.531247 4677 generic.go:334] "Generic (PLEG): container finished" podID="6c52f0ed-484d-44cf-85a3-17732f779171" containerID="738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525" exitCode=143 Dec 03 14:11:22 crc kubenswrapper[4677]: I1203 14:11:22.532155 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cf99b568-bhfkp" event={"ID":"6c52f0ed-484d-44cf-85a3-17732f779171","Type":"ContainerDied","Data":"738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525"} Dec 03 14:11:22 crc kubenswrapper[4677]: I1203 14:11:22.760824 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:11:23 crc kubenswrapper[4677]: I1203 14:11:23.548529 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0","Type":"ContainerStarted","Data":"c1cc618f54a389b5fff8cfbe8366c643b68e8167f6a951242ad0f46868903792"} Dec 03 14:11:23 crc kubenswrapper[4677]: I1203 14:11:23.583143 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Dec 03 14:11:23 crc kubenswrapper[4677]: I1203 14:11:23.808543 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-9ccdcc6d4-gw5r7" Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.339809 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.560358 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0","Type":"ContainerStarted","Data":"8062a649171f9e234463b8bb3e1098e50793e43347ad06a2bbfeed52325522b0"} Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.560402 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"80f5a100-06b3-4f3c-beb1-7df2d48b6bc0","Type":"ContainerStarted","Data":"7fa9a1f1b82aa406ed31936788c4cca5c2681abb4acfc3e4766497dcbd603acf"} Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.564981 4677 generic.go:334] "Generic (PLEG): container finished" podID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerID="bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77" exitCode=1 Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.565078 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerDied","Data":"bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77"} Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.565147 4677 scope.go:117] "RemoveContainer" containerID="2a021e5ea1a2660a7e1d09728d6977915abfabed417bd3660d03f9a97604e1f0" Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.565711 4677 scope.go:117] "RemoveContainer" containerID="bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77" Dec 03 14:11:24 crc kubenswrapper[4677]: E1203 14:11:24.565996 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f93ff7e-f6ab-4c00-8284-70a4354e576a)\"" pod="openstack/watcher-decision-engine-0" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" Dec 03 14:11:24 crc kubenswrapper[4677]: I1203 14:11:24.582903 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.582881272 podStartE2EDuration="3.582881272s" podCreationTimestamp="2025-12-03 14:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:24.579234727 +0000 UTC m=+1475.325567192" watchObservedRunningTime="2025-12-03 14:11:24.582881272 +0000 UTC m=+1475.329213727" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.029462 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5cf99b568-bhfkp" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.179:9311/healthcheck\": read tcp 10.217.0.2:41062->10.217.0.179:9311: read: connection reset by peer" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.030087 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5cf99b568-bhfkp" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.179:9311/healthcheck\": read tcp 10.217.0.2:41050->10.217.0.179:9311: read: connection reset by peer" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.533540 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.551549 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data\") pod \"6c52f0ed-484d-44cf-85a3-17732f779171\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.551613 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c52f0ed-484d-44cf-85a3-17732f779171-logs\") pod \"6c52f0ed-484d-44cf-85a3-17732f779171\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.551646 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvw5j\" (UniqueName: \"kubernetes.io/projected/6c52f0ed-484d-44cf-85a3-17732f779171-kube-api-access-bvw5j\") pod \"6c52f0ed-484d-44cf-85a3-17732f779171\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.551665 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-combined-ca-bundle\") pod \"6c52f0ed-484d-44cf-85a3-17732f779171\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.552132 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c52f0ed-484d-44cf-85a3-17732f779171-logs" (OuterVolumeSpecName: "logs") pod "6c52f0ed-484d-44cf-85a3-17732f779171" (UID: "6c52f0ed-484d-44cf-85a3-17732f779171"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.552458 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data-custom\") pod \"6c52f0ed-484d-44cf-85a3-17732f779171\" (UID: \"6c52f0ed-484d-44cf-85a3-17732f779171\") " Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.552726 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c52f0ed-484d-44cf-85a3-17732f779171-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.566573 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c52f0ed-484d-44cf-85a3-17732f779171-kube-api-access-bvw5j" (OuterVolumeSpecName: "kube-api-access-bvw5j") pod "6c52f0ed-484d-44cf-85a3-17732f779171" (UID: "6c52f0ed-484d-44cf-85a3-17732f779171"). InnerVolumeSpecName "kube-api-access-bvw5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.569970 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c52f0ed-484d-44cf-85a3-17732f779171" (UID: "6c52f0ed-484d-44cf-85a3-17732f779171"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.638877 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c52f0ed-484d-44cf-85a3-17732f779171" (UID: "6c52f0ed-484d-44cf-85a3-17732f779171"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.641855 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data" (OuterVolumeSpecName: "config-data") pod "6c52f0ed-484d-44cf-85a3-17732f779171" (UID: "6c52f0ed-484d-44cf-85a3-17732f779171"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.662717 4677 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.662760 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.662774 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvw5j\" (UniqueName: \"kubernetes.io/projected/6c52f0ed-484d-44cf-85a3-17732f779171-kube-api-access-bvw5j\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.662786 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c52f0ed-484d-44cf-85a3-17732f779171-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.674699 4677 generic.go:334] "Generic (PLEG): container finished" podID="6c52f0ed-484d-44cf-85a3-17732f779171" containerID="bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996" exitCode=0 Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.675824 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cf99b568-bhfkp" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.681331 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cf99b568-bhfkp" event={"ID":"6c52f0ed-484d-44cf-85a3-17732f779171","Type":"ContainerDied","Data":"bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996"} Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.681378 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cf99b568-bhfkp" event={"ID":"6c52f0ed-484d-44cf-85a3-17732f779171","Type":"ContainerDied","Data":"3232cdfd108c4e0352b82b175a6602a022312fd143ea2399bae6a912e93b7348"} Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.681397 4677 scope.go:117] "RemoveContainer" containerID="bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.720008 4677 scope.go:117] "RemoveContainer" containerID="738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.726257 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5cf99b568-bhfkp"] Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.734101 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5cf99b568-bhfkp"] Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.742548 4677 scope.go:117] "RemoveContainer" containerID="bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996" Dec 03 14:11:25 crc kubenswrapper[4677]: E1203 14:11:25.743528 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996\": container with ID starting with bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996 not found: ID does not exist" containerID="bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.743559 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996"} err="failed to get container status \"bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996\": rpc error: code = NotFound desc = could not find container \"bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996\": container with ID starting with bbe837b062dc9d2a180081fd989fbeb78b0738bdb3ad3d302b404c9729db2996 not found: ID does not exist" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.743577 4677 scope.go:117] "RemoveContainer" containerID="738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525" Dec 03 14:11:25 crc kubenswrapper[4677]: E1203 14:11:25.745867 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525\": container with ID starting with 738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525 not found: ID does not exist" containerID="738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.745892 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525"} err="failed to get container status \"738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525\": rpc error: code = NotFound desc = could not find container \"738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525\": container with ID starting with 738f78e4cc51d7a40ae187ba5b3abfa5d71a43ecb2207eceb4b1c4b3e9758525 not found: ID does not exist" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.871051 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85b5c78d96-8262j" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.871831 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:11:25 crc kubenswrapper[4677]: I1203 14:11:25.989736 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" path="/var/lib/kubelet/pods/6c52f0ed-484d-44cf-85a3-17732f779171/volumes" Dec 03 14:11:26 crc kubenswrapper[4677]: I1203 14:11:26.868790 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.189611 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 14:11:27 crc kubenswrapper[4677]: E1203 14:11:27.190135 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.190152 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api" Dec 03 14:11:27 crc kubenswrapper[4677]: E1203 14:11:27.190173 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api-log" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.190181 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api-log" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.190429 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api-log" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.190454 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c52f0ed-484d-44cf-85a3-17732f779171" containerName="barbican-api" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.191356 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.195673 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.200406 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-g6zsr" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.200978 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.227882 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.296435 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63acf6b9-c8d3-4e57-a08c-63dff049e551-combined-ca-bundle\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.296509 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw2dl\" (UniqueName: \"kubernetes.io/projected/63acf6b9-c8d3-4e57-a08c-63dff049e551-kube-api-access-mw2dl\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.296549 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63acf6b9-c8d3-4e57-a08c-63dff049e551-openstack-config-secret\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.297131 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63acf6b9-c8d3-4e57-a08c-63dff049e551-openstack-config\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.399740 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63acf6b9-c8d3-4e57-a08c-63dff049e551-combined-ca-bundle\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.399810 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw2dl\" (UniqueName: \"kubernetes.io/projected/63acf6b9-c8d3-4e57-a08c-63dff049e551-kube-api-access-mw2dl\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.399840 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63acf6b9-c8d3-4e57-a08c-63dff049e551-openstack-config-secret\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.399868 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63acf6b9-c8d3-4e57-a08c-63dff049e551-openstack-config\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.400677 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63acf6b9-c8d3-4e57-a08c-63dff049e551-openstack-config\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.405932 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63acf6b9-c8d3-4e57-a08c-63dff049e551-openstack-config-secret\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.406654 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63acf6b9-c8d3-4e57-a08c-63dff049e551-combined-ca-bundle\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.422281 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw2dl\" (UniqueName: \"kubernetes.io/projected/63acf6b9-c8d3-4e57-a08c-63dff049e551-kube-api-access-mw2dl\") pod \"openstackclient\" (UID: \"63acf6b9-c8d3-4e57-a08c-63dff049e551\") " pod="openstack/openstackclient" Dec 03 14:11:27 crc kubenswrapper[4677]: I1203 14:11:27.515585 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 14:11:28 crc kubenswrapper[4677]: I1203 14:11:28.023020 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 14:11:28 crc kubenswrapper[4677]: W1203 14:11:28.026677 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63acf6b9_c8d3_4e57_a08c_63dff049e551.slice/crio-040b2a02c716bc4b5a5e345315a054485eebad0a9f14149011d4d160d5dcda42 WatchSource:0}: Error finding container 040b2a02c716bc4b5a5e345315a054485eebad0a9f14149011d4d160d5dcda42: Status 404 returned error can't find the container with id 040b2a02c716bc4b5a5e345315a054485eebad0a9f14149011d4d160d5dcda42 Dec 03 14:11:28 crc kubenswrapper[4677]: I1203 14:11:28.733143 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"63acf6b9-c8d3-4e57-a08c-63dff049e551","Type":"ContainerStarted","Data":"040b2a02c716bc4b5a5e345315a054485eebad0a9f14149011d4d160d5dcda42"} Dec 03 14:11:29 crc kubenswrapper[4677]: I1203 14:11:29.160246 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:29 crc kubenswrapper[4677]: I1203 14:11:29.160316 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:29 crc kubenswrapper[4677]: I1203 14:11:29.161201 4677 scope.go:117] "RemoveContainer" containerID="bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77" Dec 03 14:11:29 crc kubenswrapper[4677]: E1203 14:11:29.161494 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f93ff7e-f6ab-4c00-8284-70a4354e576a)\"" pod="openstack/watcher-decision-engine-0" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" Dec 03 14:11:30 crc kubenswrapper[4677]: I1203 14:11:30.756293 4677 generic.go:334] "Generic (PLEG): container finished" podID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerID="e6cccf2feedc2eccb7024eb91dee2e49a9b56d56ebf604895bef4e840ca87f49" exitCode=137 Dec 03 14:11:30 crc kubenswrapper[4677]: I1203 14:11:30.756615 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b5c78d96-8262j" event={"ID":"c4ec3794-7d5b-49dc-a957-bee4761d55f0","Type":"ContainerDied","Data":"e6cccf2feedc2eccb7024eb91dee2e49a9b56d56ebf604895bef4e840ca87f49"} Dec 03 14:11:30 crc kubenswrapper[4677]: I1203 14:11:30.978409 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.175661 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-secret-key\") pod \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.176064 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-tls-certs\") pod \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.176105 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-combined-ca-bundle\") pod \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.176208 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-scripts\") pod \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.176282 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2f5n\" (UniqueName: \"kubernetes.io/projected/c4ec3794-7d5b-49dc-a957-bee4761d55f0-kube-api-access-c2f5n\") pod \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.176310 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ec3794-7d5b-49dc-a957-bee4761d55f0-logs\") pod \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.176435 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-config-data\") pod \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\" (UID: \"c4ec3794-7d5b-49dc-a957-bee4761d55f0\") " Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.176751 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ec3794-7d5b-49dc-a957-bee4761d55f0-logs" (OuterVolumeSpecName: "logs") pod "c4ec3794-7d5b-49dc-a957-bee4761d55f0" (UID: "c4ec3794-7d5b-49dc-a957-bee4761d55f0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.177100 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ec3794-7d5b-49dc-a957-bee4761d55f0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.190904 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c4ec3794-7d5b-49dc-a957-bee4761d55f0" (UID: "c4ec3794-7d5b-49dc-a957-bee4761d55f0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.191144 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ec3794-7d5b-49dc-a957-bee4761d55f0-kube-api-access-c2f5n" (OuterVolumeSpecName: "kube-api-access-c2f5n") pod "c4ec3794-7d5b-49dc-a957-bee4761d55f0" (UID: "c4ec3794-7d5b-49dc-a957-bee4761d55f0"). InnerVolumeSpecName "kube-api-access-c2f5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.216292 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-scripts" (OuterVolumeSpecName: "scripts") pod "c4ec3794-7d5b-49dc-a957-bee4761d55f0" (UID: "c4ec3794-7d5b-49dc-a957-bee4761d55f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.235175 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-config-data" (OuterVolumeSpecName: "config-data") pod "c4ec3794-7d5b-49dc-a957-bee4761d55f0" (UID: "c4ec3794-7d5b-49dc-a957-bee4761d55f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.272201 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4ec3794-7d5b-49dc-a957-bee4761d55f0" (UID: "c4ec3794-7d5b-49dc-a957-bee4761d55f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.279512 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.279556 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2f5n\" (UniqueName: \"kubernetes.io/projected/c4ec3794-7d5b-49dc-a957-bee4761d55f0-kube-api-access-c2f5n\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.279590 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ec3794-7d5b-49dc-a957-bee4761d55f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.279601 4677 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.279611 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.320854 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "c4ec3794-7d5b-49dc-a957-bee4761d55f0" (UID: "c4ec3794-7d5b-49dc-a957-bee4761d55f0"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.380978 4677 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ec3794-7d5b-49dc-a957-bee4761d55f0-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.768341 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b5c78d96-8262j" event={"ID":"c4ec3794-7d5b-49dc-a957-bee4761d55f0","Type":"ContainerDied","Data":"15153bdcffcb20da026bd3f7fcf706ce181873b5a84c6df8b05b4093603fe971"} Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.768405 4677 scope.go:117] "RemoveContainer" containerID="8a66e68be6fff49b84696df5494a916693c024f8458f6735679bc7c8ba30309a" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.768510 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b5c78d96-8262j" Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.845009 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85b5c78d96-8262j"] Dec 03 14:11:31 crc kubenswrapper[4677]: I1203 14:11:31.869221 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85b5c78d96-8262j"] Dec 03 14:11:32 crc kubenswrapper[4677]: I1203 14:11:32.017517 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" path="/var/lib/kubelet/pods/c4ec3794-7d5b-49dc-a957-bee4761d55f0/volumes" Dec 03 14:11:32 crc kubenswrapper[4677]: I1203 14:11:32.022019 4677 scope.go:117] "RemoveContainer" containerID="e6cccf2feedc2eccb7024eb91dee2e49a9b56d56ebf604895bef4e840ca87f49" Dec 03 14:11:32 crc kubenswrapper[4677]: I1203 14:11:32.297440 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.403915 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-9cc645f55-v5xrn"] Dec 03 14:11:34 crc kubenswrapper[4677]: E1203 14:11:34.407572 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon-log" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.407588 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon-log" Dec 03 14:11:34 crc kubenswrapper[4677]: E1203 14:11:34.407608 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.407614 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.407792 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon-log" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.407823 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec3794-7d5b-49dc-a957-bee4761d55f0" containerName="horizon" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.409158 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.413374 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.417995 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.418002 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.424622 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-9cc645f55-v5xrn"] Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.573210 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5da71451-5801-4ce9-b177-cd847b635e24-run-httpd\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.573293 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-config-data\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.573424 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-public-tls-certs\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.573552 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b6lv\" (UniqueName: \"kubernetes.io/projected/5da71451-5801-4ce9-b177-cd847b635e24-kube-api-access-7b6lv\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.573762 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5da71451-5801-4ce9-b177-cd847b635e24-etc-swift\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.573849 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-combined-ca-bundle\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.573898 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-internal-tls-certs\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.574009 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5da71451-5801-4ce9-b177-cd847b635e24-log-httpd\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676025 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5da71451-5801-4ce9-b177-cd847b635e24-etc-swift\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676099 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-combined-ca-bundle\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676125 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-internal-tls-certs\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676177 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5da71451-5801-4ce9-b177-cd847b635e24-log-httpd\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676240 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5da71451-5801-4ce9-b177-cd847b635e24-run-httpd\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676305 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-config-data\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676343 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-public-tls-certs\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.676377 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b6lv\" (UniqueName: \"kubernetes.io/projected/5da71451-5801-4ce9-b177-cd847b635e24-kube-api-access-7b6lv\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.677400 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5da71451-5801-4ce9-b177-cd847b635e24-log-httpd\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.677659 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5da71451-5801-4ce9-b177-cd847b635e24-run-httpd\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.684346 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5da71451-5801-4ce9-b177-cd847b635e24-etc-swift\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.685626 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-config-data\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.686177 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-internal-tls-certs\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.688707 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-combined-ca-bundle\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.712733 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da71451-5801-4ce9-b177-cd847b635e24-public-tls-certs\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.719717 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b6lv\" (UniqueName: \"kubernetes.io/projected/5da71451-5801-4ce9-b177-cd847b635e24-kube-api-access-7b6lv\") pod \"swift-proxy-9cc645f55-v5xrn\" (UID: \"5da71451-5801-4ce9-b177-cd847b635e24\") " pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:34 crc kubenswrapper[4677]: I1203 14:11:34.734394 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.027889 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.028700 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="proxy-httpd" containerID="cri-o://168f927552bf3aafd7201ec78f561e04ba823cb86d59c65a4a423d5e309d0451" gracePeriod=30 Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.028976 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-notification-agent" containerID="cri-o://5ad85e76961d44f7fe0734108239a0dc760a1819635d79bfaa56ce984ee7c150" gracePeriod=30 Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.028959 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="sg-core" containerID="cri-o://83aa0a8551635104eb88c6bbb4ba0d82ce1418736f16770c2b59eb12f5996807" gracePeriod=30 Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.029127 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-central-agent" containerID="cri-o://51481a69112e94804b51807a499203d9b51eadf16fad172012edf688f788270a" gracePeriod=30 Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.037276 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.182:3000/\": EOF" Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.817370 4677 generic.go:334] "Generic (PLEG): container finished" podID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerID="168f927552bf3aafd7201ec78f561e04ba823cb86d59c65a4a423d5e309d0451" exitCode=0 Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.817631 4677 generic.go:334] "Generic (PLEG): container finished" podID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerID="83aa0a8551635104eb88c6bbb4ba0d82ce1418736f16770c2b59eb12f5996807" exitCode=2 Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.817641 4677 generic.go:334] "Generic (PLEG): container finished" podID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerID="51481a69112e94804b51807a499203d9b51eadf16fad172012edf688f788270a" exitCode=0 Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.817661 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerDied","Data":"168f927552bf3aafd7201ec78f561e04ba823cb86d59c65a4a423d5e309d0451"} Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.817686 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerDied","Data":"83aa0a8551635104eb88c6bbb4ba0d82ce1418736f16770c2b59eb12f5996807"} Dec 03 14:11:35 crc kubenswrapper[4677]: I1203 14:11:35.817696 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerDied","Data":"51481a69112e94804b51807a499203d9b51eadf16fad172012edf688f788270a"} Dec 03 14:11:39 crc kubenswrapper[4677]: I1203 14:11:39.870064 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"63acf6b9-c8d3-4e57-a08c-63dff049e551","Type":"ContainerStarted","Data":"34af172fc7012d7e7c4eed331333d6570c43b426799a7ba203085c76d35b32f5"} Dec 03 14:11:39 crc kubenswrapper[4677]: I1203 14:11:39.872852 4677 generic.go:334] "Generic (PLEG): container finished" podID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerID="5ad85e76961d44f7fe0734108239a0dc760a1819635d79bfaa56ce984ee7c150" exitCode=0 Dec 03 14:11:39 crc kubenswrapper[4677]: I1203 14:11:39.872896 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerDied","Data":"5ad85e76961d44f7fe0734108239a0dc760a1819635d79bfaa56ce984ee7c150"} Dec 03 14:11:39 crc kubenswrapper[4677]: I1203 14:11:39.882722 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.411713687 podStartE2EDuration="12.882702133s" podCreationTimestamp="2025-12-03 14:11:27 +0000 UTC" firstStartedPulling="2025-12-03 14:11:28.029087783 +0000 UTC m=+1478.775420248" lastFinishedPulling="2025-12-03 14:11:39.500076239 +0000 UTC m=+1490.246408694" observedRunningTime="2025-12-03 14:11:39.880389502 +0000 UTC m=+1490.626721957" watchObservedRunningTime="2025-12-03 14:11:39.882702133 +0000 UTC m=+1490.629034598" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.063487 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-9cc645f55-v5xrn"] Dec 03 14:11:40 crc kubenswrapper[4677]: W1203 14:11:40.072151 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5da71451_5801_4ce9_b177_cd847b635e24.slice/crio-7aef1125c3739630e2fb481edf597bb482ef22b15dfaf7aa7156bd05ab7d0e05 WatchSource:0}: Error finding container 7aef1125c3739630e2fb481edf597bb482ef22b15dfaf7aa7156bd05ab7d0e05: Status 404 returned error can't find the container with id 7aef1125c3739630e2fb481edf597bb482ef22b15dfaf7aa7156bd05ab7d0e05 Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.116773 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.316597 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-combined-ca-bundle\") pod \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.316688 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-log-httpd\") pod \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.316777 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-sg-core-conf-yaml\") pod \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.316846 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-run-httpd\") pod \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.316924 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxmw6\" (UniqueName: \"kubernetes.io/projected/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-kube-api-access-xxmw6\") pod \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.316971 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-scripts\") pod \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.317048 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-config-data\") pod \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\" (UID: \"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b\") " Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.317323 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" (UID: "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.317360 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" (UID: "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.317838 4677 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.317867 4677 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.320893 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-scripts" (OuterVolumeSpecName: "scripts") pod "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" (UID: "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.321146 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-kube-api-access-xxmw6" (OuterVolumeSpecName: "kube-api-access-xxmw6") pod "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" (UID: "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b"). InnerVolumeSpecName "kube-api-access-xxmw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.352920 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" (UID: "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.410447 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" (UID: "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.418908 4677 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.418938 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxmw6\" (UniqueName: \"kubernetes.io/projected/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-kube-api-access-xxmw6\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.418966 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.418979 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.448240 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-config-data" (OuterVolumeSpecName: "config-data") pod "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" (UID: "257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.520219 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.885229 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9cc645f55-v5xrn" event={"ID":"5da71451-5801-4ce9-b177-cd847b635e24","Type":"ContainerStarted","Data":"7ba0e3f4d931861a3fa187102322b80ca974f36d1cba008d9ddd3fe3eae0350a"} Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.885422 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.885441 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.885449 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9cc645f55-v5xrn" event={"ID":"5da71451-5801-4ce9-b177-cd847b635e24","Type":"ContainerStarted","Data":"7b6484c5efa059fd6732c3fc06f93fcdee302b1216f3c58b4460e7cfa05ae092"} Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.885459 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9cc645f55-v5xrn" event={"ID":"5da71451-5801-4ce9-b177-cd847b635e24","Type":"ContainerStarted","Data":"7aef1125c3739630e2fb481edf597bb482ef22b15dfaf7aa7156bd05ab7d0e05"} Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.889139 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b","Type":"ContainerDied","Data":"7f57059feeda29c95bd51bb86eb93d2120e8911f2657d2673e90ba647db0afdf"} Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.889204 4677 scope.go:117] "RemoveContainer" containerID="168f927552bf3aafd7201ec78f561e04ba823cb86d59c65a4a423d5e309d0451" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.889440 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.921280 4677 scope.go:117] "RemoveContainer" containerID="83aa0a8551635104eb88c6bbb4ba0d82ce1418736f16770c2b59eb12f5996807" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.929980 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-9cc645f55-v5xrn" podStartSLOduration=6.9299610099999995 podStartE2EDuration="6.92996101s" podCreationTimestamp="2025-12-03 14:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:40.907245425 +0000 UTC m=+1491.653577900" watchObservedRunningTime="2025-12-03 14:11:40.92996101 +0000 UTC m=+1491.676293475" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.945508 4677 scope.go:117] "RemoveContainer" containerID="5ad85e76961d44f7fe0734108239a0dc760a1819635d79bfaa56ce984ee7c150" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.955814 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.966025 4677 scope.go:117] "RemoveContainer" containerID="51481a69112e94804b51807a499203d9b51eadf16fad172012edf688f788270a" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.966155 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.975544 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:40 crc kubenswrapper[4677]: E1203 14:11:40.976008 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-notification-agent" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976030 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-notification-agent" Dec 03 14:11:40 crc kubenswrapper[4677]: E1203 14:11:40.976060 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="proxy-httpd" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976067 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="proxy-httpd" Dec 03 14:11:40 crc kubenswrapper[4677]: E1203 14:11:40.976091 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="sg-core" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976099 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="sg-core" Dec 03 14:11:40 crc kubenswrapper[4677]: E1203 14:11:40.976115 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-central-agent" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976122 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-central-agent" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976359 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="proxy-httpd" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976378 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-notification-agent" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976394 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="ceilometer-central-agent" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.976410 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" containerName="sg-core" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.978565 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.980033 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.980141 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:11:40 crc kubenswrapper[4677]: I1203 14:11:40.984207 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.031553 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.031625 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-log-httpd\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.031769 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-config-data\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.031872 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwbzk\" (UniqueName: \"kubernetes.io/projected/a3227517-ddda-40a6-b219-420ca2b9a471-kube-api-access-dwbzk\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.031912 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.032025 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-run-httpd\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.032076 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-scripts\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.133480 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwbzk\" (UniqueName: \"kubernetes.io/projected/a3227517-ddda-40a6-b219-420ca2b9a471-kube-api-access-dwbzk\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.133523 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.133570 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-run-httpd\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.133591 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-scripts\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.133633 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.133650 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-log-httpd\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.133705 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-config-data\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.134627 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-run-httpd\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.138381 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-log-httpd\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.143034 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.143843 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-scripts\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.147869 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-config-data\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.148485 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.165345 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwbzk\" (UniqueName: \"kubernetes.io/projected/a3227517-ddda-40a6-b219-420ca2b9a471-kube-api-access-dwbzk\") pod \"ceilometer-0\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.303411 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.793596 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:41 crc kubenswrapper[4677]: W1203 14:11:41.838219 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice/crio-1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991 WatchSource:0}: Error finding container 1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991: Status 404 returned error can't find the container with id 1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991 Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.851630 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.912016 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerStarted","Data":"1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991"} Dec 03 14:11:41 crc kubenswrapper[4677]: I1203 14:11:41.990248 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b" path="/var/lib/kubelet/pods/257c64e3-7a6b-4d9f-8a08-6aaea8dc9f5b/volumes" Dec 03 14:11:42 crc kubenswrapper[4677]: I1203 14:11:42.960009 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerStarted","Data":"1d11c8196707560447dcb29880ce560501e30d0c4aae7f8ed8d30ebdb5c732cd"} Dec 03 14:11:42 crc kubenswrapper[4677]: I1203 14:11:42.960379 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerStarted","Data":"c9689b2052885c1df7e3667fb3c69d40866c01a4b79e4a0959a72dee81d4c294"} Dec 03 14:11:43 crc kubenswrapper[4677]: I1203 14:11:43.972052 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerStarted","Data":"85110736c893feeea6d16cc2ff499e8f25103779b5bb5cc5d6d7298a72b33651"} Dec 03 14:11:43 crc kubenswrapper[4677]: I1203 14:11:43.976248 4677 scope.go:117] "RemoveContainer" containerID="bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77" Dec 03 14:11:43 crc kubenswrapper[4677]: E1203 14:11:43.976489 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f93ff7e-f6ab-4c00-8284-70a4354e576a)\"" pod="openstack/watcher-decision-engine-0" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" Dec 03 14:11:44 crc kubenswrapper[4677]: I1203 14:11:44.985859 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerStarted","Data":"97e0e1ec4e07b3882d59baf9a60dc2eb2c0031d19b66ae85a837a39d22df18f7"} Dec 03 14:11:44 crc kubenswrapper[4677]: I1203 14:11:44.986161 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="proxy-httpd" containerID="cri-o://97e0e1ec4e07b3882d59baf9a60dc2eb2c0031d19b66ae85a837a39d22df18f7" gracePeriod=30 Dec 03 14:11:44 crc kubenswrapper[4677]: I1203 14:11:44.986124 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-notification-agent" containerID="cri-o://1d11c8196707560447dcb29880ce560501e30d0c4aae7f8ed8d30ebdb5c732cd" gracePeriod=30 Dec 03 14:11:44 crc kubenswrapper[4677]: I1203 14:11:44.986162 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="sg-core" containerID="cri-o://85110736c893feeea6d16cc2ff499e8f25103779b5bb5cc5d6d7298a72b33651" gracePeriod=30 Dec 03 14:11:44 crc kubenswrapper[4677]: I1203 14:11:44.986378 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:11:44 crc kubenswrapper[4677]: I1203 14:11:44.986047 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-central-agent" containerID="cri-o://c9689b2052885c1df7e3667fb3c69d40866c01a4b79e4a0959a72dee81d4c294" gracePeriod=30 Dec 03 14:11:45 crc kubenswrapper[4677]: I1203 14:11:45.019519 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5805649280000003 podStartE2EDuration="5.019494357s" podCreationTimestamp="2025-12-03 14:11:40 +0000 UTC" firstStartedPulling="2025-12-03 14:11:41.840482399 +0000 UTC m=+1492.586814854" lastFinishedPulling="2025-12-03 14:11:44.279411818 +0000 UTC m=+1495.025744283" observedRunningTime="2025-12-03 14:11:45.007521134 +0000 UTC m=+1495.753853599" watchObservedRunningTime="2025-12-03 14:11:45.019494357 +0000 UTC m=+1495.765826842" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.000157 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3227517-ddda-40a6-b219-420ca2b9a471" containerID="97e0e1ec4e07b3882d59baf9a60dc2eb2c0031d19b66ae85a837a39d22df18f7" exitCode=0 Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.000479 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3227517-ddda-40a6-b219-420ca2b9a471" containerID="85110736c893feeea6d16cc2ff499e8f25103779b5bb5cc5d6d7298a72b33651" exitCode=2 Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.000492 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3227517-ddda-40a6-b219-420ca2b9a471" containerID="1d11c8196707560447dcb29880ce560501e30d0c4aae7f8ed8d30ebdb5c732cd" exitCode=0 Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.000328 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerDied","Data":"97e0e1ec4e07b3882d59baf9a60dc2eb2c0031d19b66ae85a837a39d22df18f7"} Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.000534 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerDied","Data":"85110736c893feeea6d16cc2ff499e8f25103779b5bb5cc5d6d7298a72b33651"} Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.000552 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerDied","Data":"1d11c8196707560447dcb29880ce560501e30d0c4aae7f8ed8d30ebdb5c732cd"} Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.113765 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-nm2zs"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.114846 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.127827 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nm2zs"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.211369 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-csbsv"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.212711 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.231463 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-csbsv"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.267265 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-operator-scripts\") pod \"nova-api-db-create-nm2zs\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.267388 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jvzc\" (UniqueName: \"kubernetes.io/projected/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-kube-api-access-5jvzc\") pod \"nova-api-db-create-nm2zs\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.369773 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-operator-scripts\") pod \"nova-api-db-create-nm2zs\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.369827 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7q89\" (UniqueName: \"kubernetes.io/projected/5c23c782-4305-4cb7-a855-2f6dd16c45e2-kube-api-access-d7q89\") pod \"nova-cell0-db-create-csbsv\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.369908 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jvzc\" (UniqueName: \"kubernetes.io/projected/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-kube-api-access-5jvzc\") pod \"nova-api-db-create-nm2zs\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.369930 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c23c782-4305-4cb7-a855-2f6dd16c45e2-operator-scripts\") pod \"nova-cell0-db-create-csbsv\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.370635 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-operator-scripts\") pod \"nova-api-db-create-nm2zs\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.376911 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-vgwg2"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.378201 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.393135 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-27be-account-create-update-v69nt"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.402567 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.416195 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vgwg2"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.430626 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-27be-account-create-update-v69nt"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.431202 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.431413 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jvzc\" (UniqueName: \"kubernetes.io/projected/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-kube-api-access-5jvzc\") pod \"nova-api-db-create-nm2zs\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.436439 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.472384 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c23c782-4305-4cb7-a855-2f6dd16c45e2-operator-scripts\") pod \"nova-cell0-db-create-csbsv\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.472779 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7q89\" (UniqueName: \"kubernetes.io/projected/5c23c782-4305-4cb7-a855-2f6dd16c45e2-kube-api-access-d7q89\") pod \"nova-cell0-db-create-csbsv\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.474003 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c23c782-4305-4cb7-a855-2f6dd16c45e2-operator-scripts\") pod \"nova-cell0-db-create-csbsv\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.488365 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7q89\" (UniqueName: \"kubernetes.io/projected/5c23c782-4305-4cb7-a855-2f6dd16c45e2-kube-api-access-d7q89\") pod \"nova-cell0-db-create-csbsv\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.535615 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.574371 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsgvn\" (UniqueName: \"kubernetes.io/projected/23c51891-2a54-4a2b-8af7-9e18c571d3ea-kube-api-access-rsgvn\") pod \"nova-api-27be-account-create-update-v69nt\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.574434 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z7x6\" (UniqueName: \"kubernetes.io/projected/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-kube-api-access-7z7x6\") pod \"nova-cell1-db-create-vgwg2\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.574472 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-operator-scripts\") pod \"nova-cell1-db-create-vgwg2\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.574536 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23c51891-2a54-4a2b-8af7-9e18c571d3ea-operator-scripts\") pod \"nova-api-27be-account-create-update-v69nt\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.596127 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9792-account-create-update-8k85x"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.598643 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.602884 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.624380 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9792-account-create-update-8k85x"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.676523 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsgvn\" (UniqueName: \"kubernetes.io/projected/23c51891-2a54-4a2b-8af7-9e18c571d3ea-kube-api-access-rsgvn\") pod \"nova-api-27be-account-create-update-v69nt\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.678358 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z7x6\" (UniqueName: \"kubernetes.io/projected/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-kube-api-access-7z7x6\") pod \"nova-cell1-db-create-vgwg2\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.678418 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-operator-scripts\") pod \"nova-cell1-db-create-vgwg2\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.678650 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23c51891-2a54-4a2b-8af7-9e18c571d3ea-operator-scripts\") pod \"nova-api-27be-account-create-update-v69nt\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.679407 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-operator-scripts\") pod \"nova-cell1-db-create-vgwg2\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.679841 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23c51891-2a54-4a2b-8af7-9e18c571d3ea-operator-scripts\") pod \"nova-api-27be-account-create-update-v69nt\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.697409 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z7x6\" (UniqueName: \"kubernetes.io/projected/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-kube-api-access-7z7x6\") pod \"nova-cell1-db-create-vgwg2\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.699593 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsgvn\" (UniqueName: \"kubernetes.io/projected/23c51891-2a54-4a2b-8af7-9e18c571d3ea-kube-api-access-rsgvn\") pod \"nova-api-27be-account-create-update-v69nt\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.779434 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b4d9-account-create-update-dv852"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.781089 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.783633 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-operator-scripts\") pod \"nova-cell0-9792-account-create-update-8k85x\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.783777 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dfpr\" (UniqueName: \"kubernetes.io/projected/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-kube-api-access-9dfpr\") pod \"nova-cell0-9792-account-create-update-8k85x\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.785934 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.796237 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b4d9-account-create-update-dv852"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.885943 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3baef30-7171-433b-95c7-1810d7bb41a1-operator-scripts\") pod \"nova-cell1-b4d9-account-create-update-dv852\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.886013 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-operator-scripts\") pod \"nova-cell0-9792-account-create-update-8k85x\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.886157 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k62dv\" (UniqueName: \"kubernetes.io/projected/e3baef30-7171-433b-95c7-1810d7bb41a1-kube-api-access-k62dv\") pod \"nova-cell1-b4d9-account-create-update-dv852\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.886176 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dfpr\" (UniqueName: \"kubernetes.io/projected/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-kube-api-access-9dfpr\") pod \"nova-cell0-9792-account-create-update-8k85x\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.886740 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-operator-scripts\") pod \"nova-cell0-9792-account-create-update-8k85x\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.901338 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dfpr\" (UniqueName: \"kubernetes.io/projected/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-kube-api-access-9dfpr\") pod \"nova-cell0-9792-account-create-update-8k85x\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:46 crc kubenswrapper[4677]: W1203 14:11:46.977703 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice/crio-5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22 WatchSource:0}: Error finding container 5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22: Status 404 returned error can't find the container with id 5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22 Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.979052 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nm2zs"] Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.987544 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3baef30-7171-433b-95c7-1810d7bb41a1-operator-scripts\") pod \"nova-cell1-b4d9-account-create-update-dv852\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.987620 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k62dv\" (UniqueName: \"kubernetes.io/projected/e3baef30-7171-433b-95c7-1810d7bb41a1-kube-api-access-k62dv\") pod \"nova-cell1-b4d9-account-create-update-dv852\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.987980 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.988695 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3baef30-7171-433b-95c7-1810d7bb41a1-operator-scripts\") pod \"nova-cell1-b4d9-account-create-update-dv852\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.992428 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:46 crc kubenswrapper[4677]: I1203 14:11:46.996363 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.010417 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k62dv\" (UniqueName: \"kubernetes.io/projected/e3baef30-7171-433b-95c7-1810d7bb41a1-kube-api-access-k62dv\") pod \"nova-cell1-b4d9-account-create-update-dv852\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.022289 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nm2zs" event={"ID":"9d232644-1f01-4ed4-9ca5-d4259a0d1c99","Type":"ContainerStarted","Data":"5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22"} Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.113188 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.125151 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-csbsv"] Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.585181 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9792-account-create-update-8k85x"] Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.692397 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-27be-account-create-update-v69nt"] Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.700262 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vgwg2"] Dec 03 14:11:47 crc kubenswrapper[4677]: I1203 14:11:47.845682 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b4d9-account-create-update-dv852"] Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.053848 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b4d9-account-create-update-dv852" event={"ID":"e3baef30-7171-433b-95c7-1810d7bb41a1","Type":"ContainerStarted","Data":"3527c2e9a1195c9230a7215a1b153fb41ef15f1e86b084d8fe5cb9e2980177d9"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.058591 4677 generic.go:334] "Generic (PLEG): container finished" podID="5c23c782-4305-4cb7-a855-2f6dd16c45e2" containerID="fd003a296f156a154b9515e27c712505c2a2f1e48653f486ff3ea622268e3325" exitCode=0 Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.058650 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-csbsv" event={"ID":"5c23c782-4305-4cb7-a855-2f6dd16c45e2","Type":"ContainerDied","Data":"fd003a296f156a154b9515e27c712505c2a2f1e48653f486ff3ea622268e3325"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.058676 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-csbsv" event={"ID":"5c23c782-4305-4cb7-a855-2f6dd16c45e2","Type":"ContainerStarted","Data":"5497a79d2d21133304ca5aa29a9724cc90c9b72e50b8e3e397ea41ef5f9460fe"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.060104 4677 generic.go:334] "Generic (PLEG): container finished" podID="9d232644-1f01-4ed4-9ca5-d4259a0d1c99" containerID="21f066640d6b3a2cfffdaabc936b8a54af82b091cc1ed1cc841f91b39eff42f8" exitCode=0 Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.060230 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nm2zs" event={"ID":"9d232644-1f01-4ed4-9ca5-d4259a0d1c99","Type":"ContainerDied","Data":"21f066640d6b3a2cfffdaabc936b8a54af82b091cc1ed1cc841f91b39eff42f8"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.064355 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9792-account-create-update-8k85x" event={"ID":"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8","Type":"ContainerStarted","Data":"5304c636ef19ee93e710fd284c2246113003e33b289706562035a1d5ebae6bbc"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.064381 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9792-account-create-update-8k85x" event={"ID":"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8","Type":"ContainerStarted","Data":"24a44433783d5d6bea36b594cf6143d2df9cdb1d9b9c7fdfb659c8a89712a585"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.069170 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-27be-account-create-update-v69nt" event={"ID":"23c51891-2a54-4a2b-8af7-9e18c571d3ea","Type":"ContainerStarted","Data":"44b07641e6b0ad63aedc7f27be4b00c065566e9e80979d0646d57822cb1ab71c"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.069198 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-27be-account-create-update-v69nt" event={"ID":"23c51891-2a54-4a2b-8af7-9e18c571d3ea","Type":"ContainerStarted","Data":"b44a33e94ffc115d7d376d97b7284b68ced1e0382c0d8cf8951612ffa2af1960"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.072112 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vgwg2" event={"ID":"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5","Type":"ContainerStarted","Data":"79e2f298d3a112fd3e9246eb9d90433b251279eea18af696131d0c46d7baeeeb"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.072142 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vgwg2" event={"ID":"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5","Type":"ContainerStarted","Data":"f282b63e16e2807657298beaf4ff82c69d78367f6d24d2d4cf774e4196ce0eac"} Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.098133 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-27be-account-create-update-v69nt" podStartSLOduration=2.098111127 podStartE2EDuration="2.098111127s" podCreationTimestamp="2025-12-03 14:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:48.090997431 +0000 UTC m=+1498.837329896" watchObservedRunningTime="2025-12-03 14:11:48.098111127 +0000 UTC m=+1498.844443582" Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.114992 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-9792-account-create-update-8k85x" podStartSLOduration=2.114970418 podStartE2EDuration="2.114970418s" podCreationTimestamp="2025-12-03 14:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:48.1054627 +0000 UTC m=+1498.851795145" watchObservedRunningTime="2025-12-03 14:11:48.114970418 +0000 UTC m=+1498.861302893" Dec 03 14:11:48 crc kubenswrapper[4677]: I1203 14:11:48.143053 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-vgwg2" podStartSLOduration=2.143038263 podStartE2EDuration="2.143038263s" podCreationTimestamp="2025-12-03 14:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:48.140116076 +0000 UTC m=+1498.886448541" watchObservedRunningTime="2025-12-03 14:11:48.143038263 +0000 UTC m=+1498.889370718" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.089243 4677 generic.go:334] "Generic (PLEG): container finished" podID="e3baef30-7171-433b-95c7-1810d7bb41a1" containerID="06ffbe4a2a5ac7704b579d41f6e9b7ab32d7c405ec971c358f3c2dbba2c4438b" exitCode=0 Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.089354 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b4d9-account-create-update-dv852" event={"ID":"e3baef30-7171-433b-95c7-1810d7bb41a1","Type":"ContainerDied","Data":"06ffbe4a2a5ac7704b579d41f6e9b7ab32d7c405ec971c358f3c2dbba2c4438b"} Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.101174 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3227517-ddda-40a6-b219-420ca2b9a471" containerID="c9689b2052885c1df7e3667fb3c69d40866c01a4b79e4a0959a72dee81d4c294" exitCode=0 Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.101215 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerDied","Data":"c9689b2052885c1df7e3667fb3c69d40866c01a4b79e4a0959a72dee81d4c294"} Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.121656 4677 generic.go:334] "Generic (PLEG): container finished" podID="11d47538-6f9f-47ec-8d63-ddc3abb6f7a8" containerID="5304c636ef19ee93e710fd284c2246113003e33b289706562035a1d5ebae6bbc" exitCode=0 Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.121778 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9792-account-create-update-8k85x" event={"ID":"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8","Type":"ContainerDied","Data":"5304c636ef19ee93e710fd284c2246113003e33b289706562035a1d5ebae6bbc"} Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.124472 4677 generic.go:334] "Generic (PLEG): container finished" podID="23c51891-2a54-4a2b-8af7-9e18c571d3ea" containerID="44b07641e6b0ad63aedc7f27be4b00c065566e9e80979d0646d57822cb1ab71c" exitCode=0 Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.124568 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-27be-account-create-update-v69nt" event={"ID":"23c51891-2a54-4a2b-8af7-9e18c571d3ea","Type":"ContainerDied","Data":"44b07641e6b0ad63aedc7f27be4b00c065566e9e80979d0646d57822cb1ab71c"} Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.126723 4677 generic.go:334] "Generic (PLEG): container finished" podID="48e57ed4-f60c-4bd1-a9dc-68a298bd76a5" containerID="79e2f298d3a112fd3e9246eb9d90433b251279eea18af696131d0c46d7baeeeb" exitCode=0 Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.126903 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vgwg2" event={"ID":"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5","Type":"ContainerDied","Data":"79e2f298d3a112fd3e9246eb9d90433b251279eea18af696131d0c46d7baeeeb"} Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.161963 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.162997 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.163027 4677 scope.go:117] "RemoveContainer" containerID="bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.374864 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.562110 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwbzk\" (UniqueName: \"kubernetes.io/projected/a3227517-ddda-40a6-b219-420ca2b9a471-kube-api-access-dwbzk\") pod \"a3227517-ddda-40a6-b219-420ca2b9a471\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.562160 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-log-httpd\") pod \"a3227517-ddda-40a6-b219-420ca2b9a471\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.562197 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-sg-core-conf-yaml\") pod \"a3227517-ddda-40a6-b219-420ca2b9a471\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.562281 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-run-httpd\") pod \"a3227517-ddda-40a6-b219-420ca2b9a471\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.562360 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-combined-ca-bundle\") pod \"a3227517-ddda-40a6-b219-420ca2b9a471\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.562395 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-scripts\") pod \"a3227517-ddda-40a6-b219-420ca2b9a471\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.562455 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-config-data\") pod \"a3227517-ddda-40a6-b219-420ca2b9a471\" (UID: \"a3227517-ddda-40a6-b219-420ca2b9a471\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.565475 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a3227517-ddda-40a6-b219-420ca2b9a471" (UID: "a3227517-ddda-40a6-b219-420ca2b9a471"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.568821 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a3227517-ddda-40a6-b219-420ca2b9a471" (UID: "a3227517-ddda-40a6-b219-420ca2b9a471"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.572879 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-scripts" (OuterVolumeSpecName: "scripts") pod "a3227517-ddda-40a6-b219-420ca2b9a471" (UID: "a3227517-ddda-40a6-b219-420ca2b9a471"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.595459 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3227517-ddda-40a6-b219-420ca2b9a471-kube-api-access-dwbzk" (OuterVolumeSpecName: "kube-api-access-dwbzk") pod "a3227517-ddda-40a6-b219-420ca2b9a471" (UID: "a3227517-ddda-40a6-b219-420ca2b9a471"). InnerVolumeSpecName "kube-api-access-dwbzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.637831 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a3227517-ddda-40a6-b219-420ca2b9a471" (UID: "a3227517-ddda-40a6-b219-420ca2b9a471"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.664869 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.665173 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwbzk\" (UniqueName: \"kubernetes.io/projected/a3227517-ddda-40a6-b219-420ca2b9a471-kube-api-access-dwbzk\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.665238 4677 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.665349 4677 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.665418 4677 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3227517-ddda-40a6-b219-420ca2b9a471-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.691729 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-config-data" (OuterVolumeSpecName: "config-data") pod "a3227517-ddda-40a6-b219-420ca2b9a471" (UID: "a3227517-ddda-40a6-b219-420ca2b9a471"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.699285 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3227517-ddda-40a6-b219-420ca2b9a471" (UID: "a3227517-ddda-40a6-b219-420ca2b9a471"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.742549 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.749456 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-9cc645f55-v5xrn" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.767463 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.767500 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3227517-ddda-40a6-b219-420ca2b9a471-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.781442 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.818936 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.868489 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c23c782-4305-4cb7-a855-2f6dd16c45e2-operator-scripts\") pod \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.868640 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7q89\" (UniqueName: \"kubernetes.io/projected/5c23c782-4305-4cb7-a855-2f6dd16c45e2-kube-api-access-d7q89\") pod \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\" (UID: \"5c23c782-4305-4cb7-a855-2f6dd16c45e2\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.869320 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c23c782-4305-4cb7-a855-2f6dd16c45e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c23c782-4305-4cb7-a855-2f6dd16c45e2" (UID: "5c23c782-4305-4cb7-a855-2f6dd16c45e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.875168 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c23c782-4305-4cb7-a855-2f6dd16c45e2-kube-api-access-d7q89" (OuterVolumeSpecName: "kube-api-access-d7q89") pod "5c23c782-4305-4cb7-a855-2f6dd16c45e2" (UID: "5c23c782-4305-4cb7-a855-2f6dd16c45e2"). InnerVolumeSpecName "kube-api-access-d7q89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.970672 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jvzc\" (UniqueName: \"kubernetes.io/projected/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-kube-api-access-5jvzc\") pod \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.970721 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-operator-scripts\") pod \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\" (UID: \"9d232644-1f01-4ed4-9ca5-d4259a0d1c99\") " Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.971184 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d232644-1f01-4ed4-9ca5-d4259a0d1c99" (UID: "9d232644-1f01-4ed4-9ca5-d4259a0d1c99"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.971198 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7q89\" (UniqueName: \"kubernetes.io/projected/5c23c782-4305-4cb7-a855-2f6dd16c45e2-kube-api-access-d7q89\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.971363 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c23c782-4305-4cb7-a855-2f6dd16c45e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:49 crc kubenswrapper[4677]: I1203 14:11:49.974336 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-kube-api-access-5jvzc" (OuterVolumeSpecName: "kube-api-access-5jvzc") pod "9d232644-1f01-4ed4-9ca5-d4259a0d1c99" (UID: "9d232644-1f01-4ed4-9ca5-d4259a0d1c99"). InnerVolumeSpecName "kube-api-access-5jvzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.074717 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jvzc\" (UniqueName: \"kubernetes.io/projected/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-kube-api-access-5jvzc\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.074751 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d232644-1f01-4ed4-9ca5-d4259a0d1c99-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.137162 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3227517-ddda-40a6-b219-420ca2b9a471","Type":"ContainerDied","Data":"1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991"} Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.137212 4677 scope.go:117] "RemoveContainer" containerID="97e0e1ec4e07b3882d59baf9a60dc2eb2c0031d19b66ae85a837a39d22df18f7" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.137341 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.142732 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nm2zs" event={"ID":"9d232644-1f01-4ed4-9ca5-d4259a0d1c99","Type":"ContainerDied","Data":"5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22"} Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.142761 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.142803 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nm2zs" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.148271 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-csbsv" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.149363 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-csbsv" event={"ID":"5c23c782-4305-4cb7-a855-2f6dd16c45e2","Type":"ContainerDied","Data":"5497a79d2d21133304ca5aa29a9724cc90c9b72e50b8e3e397ea41ef5f9460fe"} Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.149396 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5497a79d2d21133304ca5aa29a9724cc90c9b72e50b8e3e397ea41ef5f9460fe" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.162742 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.181983 4677 scope.go:117] "RemoveContainer" containerID="85110736c893feeea6d16cc2ff499e8f25103779b5bb5cc5d6d7298a72b33651" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.187341 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.197526 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:50 crc kubenswrapper[4677]: E1203 14:11:50.198121 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d232644-1f01-4ed4-9ca5-d4259a0d1c99" containerName="mariadb-database-create" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198146 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d232644-1f01-4ed4-9ca5-d4259a0d1c99" containerName="mariadb-database-create" Dec 03 14:11:50 crc kubenswrapper[4677]: E1203 14:11:50.198165 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-notification-agent" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198173 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-notification-agent" Dec 03 14:11:50 crc kubenswrapper[4677]: E1203 14:11:50.198192 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c23c782-4305-4cb7-a855-2f6dd16c45e2" containerName="mariadb-database-create" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198200 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c23c782-4305-4cb7-a855-2f6dd16c45e2" containerName="mariadb-database-create" Dec 03 14:11:50 crc kubenswrapper[4677]: E1203 14:11:50.198223 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="proxy-httpd" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198233 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="proxy-httpd" Dec 03 14:11:50 crc kubenswrapper[4677]: E1203 14:11:50.198247 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-central-agent" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198255 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-central-agent" Dec 03 14:11:50 crc kubenswrapper[4677]: E1203 14:11:50.198276 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="sg-core" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198283 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="sg-core" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198523 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="sg-core" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198543 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d232644-1f01-4ed4-9ca5-d4259a0d1c99" containerName="mariadb-database-create" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198571 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-central-agent" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198583 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="proxy-httpd" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198595 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" containerName="ceilometer-notification-agent" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.198604 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c23c782-4305-4cb7-a855-2f6dd16c45e2" containerName="mariadb-database-create" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.200683 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.203885 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.204218 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.225287 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.381968 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pj4q\" (UniqueName: \"kubernetes.io/projected/8322e4ce-8390-4013-a3e2-d264e5f7568f-kube-api-access-5pj4q\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.382133 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-config-data\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.382185 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-log-httpd\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.382319 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-scripts\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.382492 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.382548 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-run-httpd\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.382656 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.484256 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-run-httpd\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.484342 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.484405 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pj4q\" (UniqueName: \"kubernetes.io/projected/8322e4ce-8390-4013-a3e2-d264e5f7568f-kube-api-access-5pj4q\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.484431 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-config-data\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.484453 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-log-httpd\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.484470 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-scripts\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.484512 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.485678 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-log-httpd\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.486546 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-run-httpd\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.493857 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.494940 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.496127 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-config-data\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.500845 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-scripts\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.536127 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pj4q\" (UniqueName: \"kubernetes.io/projected/8322e4ce-8390-4013-a3e2-d264e5f7568f-kube-api-access-5pj4q\") pod \"ceilometer-0\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.561057 4677 scope.go:117] "RemoveContainer" containerID="1d11c8196707560447dcb29880ce560501e30d0c4aae7f8ed8d30ebdb5c732cd" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.624348 4677 scope.go:117] "RemoveContainer" containerID="c9689b2052885c1df7e3667fb3c69d40866c01a4b79e4a0959a72dee81d4c294" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.737127 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.755569 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.829866 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.890745 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k62dv\" (UniqueName: \"kubernetes.io/projected/e3baef30-7171-433b-95c7-1810d7bb41a1-kube-api-access-k62dv\") pod \"e3baef30-7171-433b-95c7-1810d7bb41a1\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.890933 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-operator-scripts\") pod \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.891083 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3baef30-7171-433b-95c7-1810d7bb41a1-operator-scripts\") pod \"e3baef30-7171-433b-95c7-1810d7bb41a1\" (UID: \"e3baef30-7171-433b-95c7-1810d7bb41a1\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.891199 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dfpr\" (UniqueName: \"kubernetes.io/projected/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-kube-api-access-9dfpr\") pod \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\" (UID: \"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.892661 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3baef30-7171-433b-95c7-1810d7bb41a1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e3baef30-7171-433b-95c7-1810d7bb41a1" (UID: "e3baef30-7171-433b-95c7-1810d7bb41a1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.892678 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11d47538-6f9f-47ec-8d63-ddc3abb6f7a8" (UID: "11d47538-6f9f-47ec-8d63-ddc3abb6f7a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.896816 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3baef30-7171-433b-95c7-1810d7bb41a1-kube-api-access-k62dv" (OuterVolumeSpecName: "kube-api-access-k62dv") pod "e3baef30-7171-433b-95c7-1810d7bb41a1" (UID: "e3baef30-7171-433b-95c7-1810d7bb41a1"). InnerVolumeSpecName "kube-api-access-k62dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.899451 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-kube-api-access-9dfpr" (OuterVolumeSpecName: "kube-api-access-9dfpr") pod "11d47538-6f9f-47ec-8d63-ddc3abb6f7a8" (UID: "11d47538-6f9f-47ec-8d63-ddc3abb6f7a8"). InnerVolumeSpecName "kube-api-access-9dfpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.904560 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.909465 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.992851 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsgvn\" (UniqueName: \"kubernetes.io/projected/23c51891-2a54-4a2b-8af7-9e18c571d3ea-kube-api-access-rsgvn\") pod \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.993003 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23c51891-2a54-4a2b-8af7-9e18c571d3ea-operator-scripts\") pod \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\" (UID: \"23c51891-2a54-4a2b-8af7-9e18c571d3ea\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.993036 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z7x6\" (UniqueName: \"kubernetes.io/projected/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-kube-api-access-7z7x6\") pod \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.993133 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-operator-scripts\") pod \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\" (UID: \"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5\") " Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.993534 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dfpr\" (UniqueName: \"kubernetes.io/projected/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-kube-api-access-9dfpr\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.993546 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k62dv\" (UniqueName: \"kubernetes.io/projected/e3baef30-7171-433b-95c7-1810d7bb41a1-kube-api-access-k62dv\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.993556 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.993564 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3baef30-7171-433b-95c7-1810d7bb41a1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.994078 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23c51891-2a54-4a2b-8af7-9e18c571d3ea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "23c51891-2a54-4a2b-8af7-9e18c571d3ea" (UID: "23c51891-2a54-4a2b-8af7-9e18c571d3ea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.994162 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48e57ed4-f60c-4bd1-a9dc-68a298bd76a5" (UID: "48e57ed4-f60c-4bd1-a9dc-68a298bd76a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.997402 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-kube-api-access-7z7x6" (OuterVolumeSpecName: "kube-api-access-7z7x6") pod "48e57ed4-f60c-4bd1-a9dc-68a298bd76a5" (UID: "48e57ed4-f60c-4bd1-a9dc-68a298bd76a5"). InnerVolumeSpecName "kube-api-access-7z7x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:50 crc kubenswrapper[4677]: I1203 14:11:50.997429 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23c51891-2a54-4a2b-8af7-9e18c571d3ea-kube-api-access-rsgvn" (OuterVolumeSpecName: "kube-api-access-rsgvn") pod "23c51891-2a54-4a2b-8af7-9e18c571d3ea" (UID: "23c51891-2a54-4a2b-8af7-9e18c571d3ea"). InnerVolumeSpecName "kube-api-access-rsgvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.101509 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.101538 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsgvn\" (UniqueName: \"kubernetes.io/projected/23c51891-2a54-4a2b-8af7-9e18c571d3ea-kube-api-access-rsgvn\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.101549 4677 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23c51891-2a54-4a2b-8af7-9e18c571d3ea-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.101558 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z7x6\" (UniqueName: \"kubernetes.io/projected/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5-kube-api-access-7z7x6\") on node \"crc\" DevicePath \"\"" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.165108 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-27be-account-create-update-v69nt" event={"ID":"23c51891-2a54-4a2b-8af7-9e18c571d3ea","Type":"ContainerDied","Data":"b44a33e94ffc115d7d376d97b7284b68ced1e0382c0d8cf8951612ffa2af1960"} Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.165404 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b44a33e94ffc115d7d376d97b7284b68ced1e0382c0d8cf8951612ffa2af1960" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.165136 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-27be-account-create-update-v69nt" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.166917 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vgwg2" event={"ID":"48e57ed4-f60c-4bd1-a9dc-68a298bd76a5","Type":"ContainerDied","Data":"f282b63e16e2807657298beaf4ff82c69d78367f6d24d2d4cf774e4196ce0eac"} Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.166972 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f282b63e16e2807657298beaf4ff82c69d78367f6d24d2d4cf774e4196ce0eac" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.166978 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vgwg2" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.169240 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b4d9-account-create-update-dv852" event={"ID":"e3baef30-7171-433b-95c7-1810d7bb41a1","Type":"ContainerDied","Data":"3527c2e9a1195c9230a7215a1b153fb41ef15f1e86b084d8fe5cb9e2980177d9"} Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.169267 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3527c2e9a1195c9230a7215a1b153fb41ef15f1e86b084d8fe5cb9e2980177d9" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.169314 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b4d9-account-create-update-dv852" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.196556 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9792-account-create-update-8k85x" event={"ID":"11d47538-6f9f-47ec-8d63-ddc3abb6f7a8","Type":"ContainerDied","Data":"24a44433783d5d6bea36b594cf6143d2df9cdb1d9b9c7fdfb659c8a89712a585"} Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.196626 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24a44433783d5d6bea36b594cf6143d2df9cdb1d9b9c7fdfb659c8a89712a585" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.196711 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9792-account-create-update-8k85x" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.297228 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:51 crc kubenswrapper[4677]: W1203 14:11:51.297812 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8322e4ce_8390_4013_a3e2_d264e5f7568f.slice/crio-ca9db80aebf2eb7556caceea35b5201b3c76b5f76ff5fdebbe92b3c86e7c09f5 WatchSource:0}: Error finding container ca9db80aebf2eb7556caceea35b5201b3c76b5f76ff5fdebbe92b3c86e7c09f5: Status 404 returned error can't find the container with id ca9db80aebf2eb7556caceea35b5201b3c76b5f76ff5fdebbe92b3c86e7c09f5 Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.951397 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkxw5"] Dec 03 14:11:51 crc kubenswrapper[4677]: E1203 14:11:51.952072 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3baef30-7171-433b-95c7-1810d7bb41a1" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952091 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3baef30-7171-433b-95c7-1810d7bb41a1" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: E1203 14:11:51.952115 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23c51891-2a54-4a2b-8af7-9e18c571d3ea" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952122 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="23c51891-2a54-4a2b-8af7-9e18c571d3ea" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: E1203 14:11:51.952131 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e57ed4-f60c-4bd1-a9dc-68a298bd76a5" containerName="mariadb-database-create" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952137 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e57ed4-f60c-4bd1-a9dc-68a298bd76a5" containerName="mariadb-database-create" Dec 03 14:11:51 crc kubenswrapper[4677]: E1203 14:11:51.952147 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11d47538-6f9f-47ec-8d63-ddc3abb6f7a8" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952154 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="11d47538-6f9f-47ec-8d63-ddc3abb6f7a8" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952321 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="11d47538-6f9f-47ec-8d63-ddc3abb6f7a8" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952337 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e57ed4-f60c-4bd1-a9dc-68a298bd76a5" containerName="mariadb-database-create" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952356 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3baef30-7171-433b-95c7-1810d7bb41a1" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952372 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="23c51891-2a54-4a2b-8af7-9e18c571d3ea" containerName="mariadb-account-create-update" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.952987 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:51 crc kubenswrapper[4677]: E1203 14:11:51.960797 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice/crio-1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice/crio-5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22\": RecentStats: unable to find data in memory cache]" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.969852 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkxw5"] Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.970623 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.970885 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.973197 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s7d94" Dec 03 14:11:51 crc kubenswrapper[4677]: I1203 14:11:51.989133 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3227517-ddda-40a6-b219-420ca2b9a471" path="/var/lib/kubelet/pods/a3227517-ddda-40a6-b219-420ca2b9a471/volumes" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.126687 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-scripts\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.126915 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.127114 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtdg5\" (UniqueName: \"kubernetes.io/projected/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-kube-api-access-wtdg5\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.127191 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-config-data\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.208270 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerStarted","Data":"8eb03e8a7553e99a3829b7daba8b1f6493a8079ffce14c6efcbddcd626a9561f"} Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.208313 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerStarted","Data":"23912dda68269bd54ba8dba59d09c5050701f64f121779ddc45dc2e6eab3bb68"} Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.208321 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerStarted","Data":"ca9db80aebf2eb7556caceea35b5201b3c76b5f76ff5fdebbe92b3c86e7c09f5"} Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.209810 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerStarted","Data":"4308dd27ba345528b74a491ff2f517c91641232e620432c6ea4b6271efa48e9b"} Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.228532 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.228600 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtdg5\" (UniqueName: \"kubernetes.io/projected/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-kube-api-access-wtdg5\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.228632 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-config-data\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.228672 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-scripts\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.232454 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-scripts\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.233125 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.235574 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-config-data\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.263521 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtdg5\" (UniqueName: \"kubernetes.io/projected/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-kube-api-access-wtdg5\") pod \"nova-cell0-conductor-db-sync-nkxw5\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.290255 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:11:52 crc kubenswrapper[4677]: I1203 14:11:52.804366 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkxw5"] Dec 03 14:11:53 crc kubenswrapper[4677]: I1203 14:11:53.229912 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" event={"ID":"b9b349d0-d7aa-45cb-aab4-d50379b27c4d","Type":"ContainerStarted","Data":"3341af9480ba9a4a35e454f1f41316beb7b71c42474f16bf12c4d3ee22eac172"} Dec 03 14:11:54 crc kubenswrapper[4677]: I1203 14:11:54.249715 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerStarted","Data":"a7f778ef044e0ed9a61d6efa1f1db0f5af8e0174c2f666412f899bc8e5875a7f"} Dec 03 14:11:55 crc kubenswrapper[4677]: I1203 14:11:55.386689 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:11:55 crc kubenswrapper[4677]: I1203 14:11:55.387238 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-log" containerID="cri-o://e86632b6cc6cc51b64f89d50ed680cd0451cc04ee86ae40f18ab50470fc6ffef" gracePeriod=30 Dec 03 14:11:55 crc kubenswrapper[4677]: I1203 14:11:55.387341 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-httpd" containerID="cri-o://10245f6819d6abec12abcada82177e8e22fed9a10dafe05c55136c7f886838f9" gracePeriod=30 Dec 03 14:11:56 crc kubenswrapper[4677]: I1203 14:11:56.278070 4677 generic.go:334] "Generic (PLEG): container finished" podID="6945799b-1144-45b8-be28-8de8ad7b257e" containerID="e86632b6cc6cc51b64f89d50ed680cd0451cc04ee86ae40f18ab50470fc6ffef" exitCode=143 Dec 03 14:11:56 crc kubenswrapper[4677]: I1203 14:11:56.278368 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6945799b-1144-45b8-be28-8de8ad7b257e","Type":"ContainerDied","Data":"e86632b6cc6cc51b64f89d50ed680cd0451cc04ee86ae40f18ab50470fc6ffef"} Dec 03 14:11:57 crc kubenswrapper[4677]: I1203 14:11:57.307056 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerStarted","Data":"7d23fce7522889cfbe7adc1630137d136863fbc7c656995a2e903377d2616952"} Dec 03 14:11:57 crc kubenswrapper[4677]: I1203 14:11:57.307644 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:11:57 crc kubenswrapper[4677]: I1203 14:11:57.312999 4677 generic.go:334] "Generic (PLEG): container finished" podID="6945799b-1144-45b8-be28-8de8ad7b257e" containerID="10245f6819d6abec12abcada82177e8e22fed9a10dafe05c55136c7f886838f9" exitCode=0 Dec 03 14:11:57 crc kubenswrapper[4677]: I1203 14:11:57.313040 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6945799b-1144-45b8-be28-8de8ad7b257e","Type":"ContainerDied","Data":"10245f6819d6abec12abcada82177e8e22fed9a10dafe05c55136c7f886838f9"} Dec 03 14:11:57 crc kubenswrapper[4677]: I1203 14:11:57.335722 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.274017715 podStartE2EDuration="7.335701063s" podCreationTimestamp="2025-12-03 14:11:50 +0000 UTC" firstStartedPulling="2025-12-03 14:11:51.30063481 +0000 UTC m=+1502.046967265" lastFinishedPulling="2025-12-03 14:11:56.362318158 +0000 UTC m=+1507.108650613" observedRunningTime="2025-12-03 14:11:57.32376971 +0000 UTC m=+1508.070102175" watchObservedRunningTime="2025-12-03 14:11:57.335701063 +0000 UTC m=+1508.082033518" Dec 03 14:11:57 crc kubenswrapper[4677]: I1203 14:11:57.601694 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.435749 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x4m64"] Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.438768 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.450632 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x4m64"] Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.566893 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-catalog-content\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.567054 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-utilities\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.567156 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhgmk\" (UniqueName: \"kubernetes.io/projected/e2dbbddf-0907-479f-a696-f5fecd15c4da-kube-api-access-vhgmk\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.668752 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhgmk\" (UniqueName: \"kubernetes.io/projected/e2dbbddf-0907-479f-a696-f5fecd15c4da-kube-api-access-vhgmk\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.668916 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-catalog-content\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.668942 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-utilities\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.669520 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-utilities\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.669557 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-catalog-content\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.686747 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhgmk\" (UniqueName: \"kubernetes.io/projected/e2dbbddf-0907-479f-a696-f5fecd15c4da-kube-api-access-vhgmk\") pod \"redhat-operators-x4m64\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:58 crc kubenswrapper[4677]: I1203 14:11:58.764681 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.160561 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.198582 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.337494 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.337748 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-log" containerID="cri-o://76bf2f570e3bbbcf5bc48c476c42d98548b1d6d49fcb95f4d041d75dbed406f3" gracePeriod=30 Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.337870 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-httpd" containerID="cri-o://6510391c92397ab7f26b1567ed8a627f449c9ce0cae708a46a17278bbb069e05" gracePeriod=30 Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.351636 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.351871 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-central-agent" containerID="cri-o://23912dda68269bd54ba8dba59d09c5050701f64f121779ddc45dc2e6eab3bb68" gracePeriod=30 Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.352089 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="sg-core" containerID="cri-o://a7f778ef044e0ed9a61d6efa1f1db0f5af8e0174c2f666412f899bc8e5875a7f" gracePeriod=30 Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.352107 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="proxy-httpd" containerID="cri-o://7d23fce7522889cfbe7adc1630137d136863fbc7c656995a2e903377d2616952" gracePeriod=30 Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.352094 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-notification-agent" containerID="cri-o://8eb03e8a7553e99a3829b7daba8b1f6493a8079ffce14c6efcbddcd626a9561f" gracePeriod=30 Dec 03 14:11:59 crc kubenswrapper[4677]: I1203 14:11:59.408824 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375422 4677 generic.go:334] "Generic (PLEG): container finished" podID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerID="7d23fce7522889cfbe7adc1630137d136863fbc7c656995a2e903377d2616952" exitCode=0 Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375712 4677 generic.go:334] "Generic (PLEG): container finished" podID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerID="a7f778ef044e0ed9a61d6efa1f1db0f5af8e0174c2f666412f899bc8e5875a7f" exitCode=2 Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375726 4677 generic.go:334] "Generic (PLEG): container finished" podID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerID="8eb03e8a7553e99a3829b7daba8b1f6493a8079ffce14c6efcbddcd626a9561f" exitCode=0 Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375734 4677 generic.go:334] "Generic (PLEG): container finished" podID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerID="23912dda68269bd54ba8dba59d09c5050701f64f121779ddc45dc2e6eab3bb68" exitCode=0 Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375649 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerDied","Data":"7d23fce7522889cfbe7adc1630137d136863fbc7c656995a2e903377d2616952"} Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375808 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerDied","Data":"a7f778ef044e0ed9a61d6efa1f1db0f5af8e0174c2f666412f899bc8e5875a7f"} Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375827 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerDied","Data":"8eb03e8a7553e99a3829b7daba8b1f6493a8079ffce14c6efcbddcd626a9561f"} Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.375841 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerDied","Data":"23912dda68269bd54ba8dba59d09c5050701f64f121779ddc45dc2e6eab3bb68"} Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.378087 4677 generic.go:334] "Generic (PLEG): container finished" podID="232a4115-1312-485b-8df6-cad4097e7b6f" containerID="76bf2f570e3bbbcf5bc48c476c42d98548b1d6d49fcb95f4d041d75dbed406f3" exitCode=143 Dec 03 14:12:00 crc kubenswrapper[4677]: I1203 14:12:00.378345 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"232a4115-1312-485b-8df6-cad4097e7b6f","Type":"ContainerDied","Data":"76bf2f570e3bbbcf5bc48c476c42d98548b1d6d49fcb95f4d041d75dbed406f3"} Dec 03 14:12:01 crc kubenswrapper[4677]: I1203 14:12:01.389341 4677 generic.go:334] "Generic (PLEG): container finished" podID="232a4115-1312-485b-8df6-cad4097e7b6f" containerID="6510391c92397ab7f26b1567ed8a627f449c9ce0cae708a46a17278bbb069e05" exitCode=0 Dec 03 14:12:01 crc kubenswrapper[4677]: I1203 14:12:01.389416 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"232a4115-1312-485b-8df6-cad4097e7b6f","Type":"ContainerDied","Data":"6510391c92397ab7f26b1567ed8a627f449c9ce0cae708a46a17278bbb069e05"} Dec 03 14:12:02 crc kubenswrapper[4677]: E1203 14:12:02.235641 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice/crio-1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice/crio-5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22\": RecentStats: unable to find data in memory cache]" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.272425 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.371679 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-httpd-run\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372061 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-combined-ca-bundle\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372092 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372385 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-scripts\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372471 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372621 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-public-tls-certs\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372736 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-config-data\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372962 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-logs\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.372987 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhtgl\" (UniqueName: \"kubernetes.io/projected/6945799b-1144-45b8-be28-8de8ad7b257e-kube-api-access-mhtgl\") pod \"6945799b-1144-45b8-be28-8de8ad7b257e\" (UID: \"6945799b-1144-45b8-be28-8de8ad7b257e\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.373437 4677 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.376760 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-logs" (OuterVolumeSpecName: "logs") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.378801 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.381916 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-scripts" (OuterVolumeSpecName: "scripts") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.402668 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6945799b-1144-45b8-be28-8de8ad7b257e-kube-api-access-mhtgl" (OuterVolumeSpecName: "kube-api-access-mhtgl") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "kube-api-access-mhtgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.415107 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6945799b-1144-45b8-be28-8de8ad7b257e","Type":"ContainerDied","Data":"46d8fa12adb8d86245b5f61f8e2d04381a0bbc48cf74a64fccfaf1a5ff5e3985"} Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.415377 4677 scope.go:117] "RemoveContainer" containerID="10245f6819d6abec12abcada82177e8e22fed9a10dafe05c55136c7f886838f9" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.415508 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.421576 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.463830 4677 scope.go:117] "RemoveContainer" containerID="e86632b6cc6cc51b64f89d50ed680cd0451cc04ee86ae40f18ab50470fc6ffef" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.478270 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6945799b-1144-45b8-be28-8de8ad7b257e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.478289 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhtgl\" (UniqueName: \"kubernetes.io/projected/6945799b-1144-45b8-be28-8de8ad7b257e-kube-api-access-mhtgl\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.478304 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.478333 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.478346 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.480704 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-config-data" (OuterVolumeSpecName: "config-data") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.525102 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6945799b-1144-45b8-be28-8de8ad7b257e" (UID: "6945799b-1144-45b8-be28-8de8ad7b257e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.531577 4677 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.581414 4677 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.581447 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6945799b-1144-45b8-be28-8de8ad7b257e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.581457 4677 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.670434 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.705511 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.784431 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785040 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-sg-core-conf-yaml\") pod \"8322e4ce-8390-4013-a3e2-d264e5f7568f\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785104 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-scripts\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785137 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-log-httpd\") pod \"8322e4ce-8390-4013-a3e2-d264e5f7568f\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785209 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-httpd-run\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785230 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-run-httpd\") pod \"8322e4ce-8390-4013-a3e2-d264e5f7568f\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785279 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-scripts\") pod \"8322e4ce-8390-4013-a3e2-d264e5f7568f\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785315 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-combined-ca-bundle\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785361 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785426 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btzns\" (UniqueName: \"kubernetes.io/projected/232a4115-1312-485b-8df6-cad4097e7b6f-kube-api-access-btzns\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785482 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-combined-ca-bundle\") pod \"8322e4ce-8390-4013-a3e2-d264e5f7568f\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785511 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-config-data\") pod \"8322e4ce-8390-4013-a3e2-d264e5f7568f\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785530 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-logs\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785550 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-internal-tls-certs\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785600 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pj4q\" (UniqueName: \"kubernetes.io/projected/8322e4ce-8390-4013-a3e2-d264e5f7568f-kube-api-access-5pj4q\") pod \"8322e4ce-8390-4013-a3e2-d264e5f7568f\" (UID: \"8322e4ce-8390-4013-a3e2-d264e5f7568f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.785618 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-config-data\") pod \"232a4115-1312-485b-8df6-cad4097e7b6f\" (UID: \"232a4115-1312-485b-8df6-cad4097e7b6f\") " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.786703 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8322e4ce-8390-4013-a3e2-d264e5f7568f" (UID: "8322e4ce-8390-4013-a3e2-d264e5f7568f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.787139 4677 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.788385 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-logs" (OuterVolumeSpecName: "logs") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.789614 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8322e4ce-8390-4013-a3e2-d264e5f7568f" (UID: "8322e4ce-8390-4013-a3e2-d264e5f7568f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.789871 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.793515 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.803982 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8322e4ce-8390-4013-a3e2-d264e5f7568f-kube-api-access-5pj4q" (OuterVolumeSpecName: "kube-api-access-5pj4q") pod "8322e4ce-8390-4013-a3e2-d264e5f7568f" (UID: "8322e4ce-8390-4013-a3e2-d264e5f7568f"). InnerVolumeSpecName "kube-api-access-5pj4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.805873 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-scripts" (OuterVolumeSpecName: "scripts") pod "8322e4ce-8390-4013-a3e2-d264e5f7568f" (UID: "8322e4ce-8390-4013-a3e2-d264e5f7568f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.807226 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.808102 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-scripts" (OuterVolumeSpecName: "scripts") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.823373 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.823867 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="proxy-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.823960 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="proxy-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.824037 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-central-agent" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.824138 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-central-agent" Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.824219 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-log" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.824271 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-log" Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.824340 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-notification-agent" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.824420 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-notification-agent" Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.824484 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.824539 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.824650 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-log" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.824701 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-log" Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.824760 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.824822 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: E1203 14:12:03.824887 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="sg-core" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.824941 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="sg-core" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825182 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-log" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825246 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-central-agent" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825324 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-log" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825400 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="sg-core" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825457 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="ceilometer-notification-agent" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825516 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" containerName="glance-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825567 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" containerName="glance-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.825633 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" containerName="proxy-httpd" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.826819 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.832119 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232a4115-1312-485b-8df6-cad4097e7b6f-kube-api-access-btzns" (OuterVolumeSpecName: "kube-api-access-btzns") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "kube-api-access-btzns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.832524 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.834204 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.837194 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.847409 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x4m64"] Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.870296 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8322e4ce-8390-4013-a3e2-d264e5f7568f" (UID: "8322e4ce-8390-4013-a3e2-d264e5f7568f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892003 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btzns\" (UniqueName: \"kubernetes.io/projected/232a4115-1312-485b-8df6-cad4097e7b6f-kube-api-access-btzns\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892045 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892055 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pj4q\" (UniqueName: \"kubernetes.io/projected/8322e4ce-8390-4013-a3e2-d264e5f7568f-kube-api-access-5pj4q\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892066 4677 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892074 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892082 4677 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/232a4115-1312-485b-8df6-cad4097e7b6f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892090 4677 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8322e4ce-8390-4013-a3e2-d264e5f7568f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892098 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.892116 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.913877 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.969136 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.979023 4677 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.984491 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-config-data" (OuterVolumeSpecName: "config-data") pod "232a4115-1312-485b-8df6-cad4097e7b6f" (UID: "232a4115-1312-485b-8df6-cad4097e7b6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.987781 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8322e4ce-8390-4013-a3e2-d264e5f7568f" (UID: "8322e4ce-8390-4013-a3e2-d264e5f7568f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.992190 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6945799b-1144-45b8-be28-8de8ad7b257e" path="/var/lib/kubelet/pods/6945799b-1144-45b8-be28-8de8ad7b257e/volumes" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.994216 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.994279 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.994384 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.994568 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-logs\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.994813 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995073 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995161 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnr24\" (UniqueName: \"kubernetes.io/projected/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-kube-api-access-lnr24\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995196 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995483 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995499 4677 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995520 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995531 4677 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:03 crc kubenswrapper[4677]: I1203 14:12:03.995540 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232a4115-1312-485b-8df6-cad4097e7b6f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.036923 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-config-data" (OuterVolumeSpecName: "config-data") pod "8322e4ce-8390-4013-a3e2-d264e5f7568f" (UID: "8322e4ce-8390-4013-a3e2-d264e5f7568f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.098152 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.098234 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnr24\" (UniqueName: \"kubernetes.io/projected/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-kube-api-access-lnr24\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.098576 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.098606 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.098874 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.098939 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.099016 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.099080 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-logs\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.099168 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.099560 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322e4ce-8390-4013-a3e2-d264e5f7568f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.099873 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.100660 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-logs\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.102835 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.104246 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.104533 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.105765 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.159250 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnr24\" (UniqueName: \"kubernetes.io/projected/23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c-kube-api-access-lnr24\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.162084 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.221983 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.429095 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" event={"ID":"b9b349d0-d7aa-45cb-aab4-d50379b27c4d","Type":"ContainerStarted","Data":"047d9dc6493b6bc1a221b6c2bdad811a81529f68519c088a20b717a7fd6bbd15"} Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.432996 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"232a4115-1312-485b-8df6-cad4097e7b6f","Type":"ContainerDied","Data":"6e3e15a73c88cd2288469c9158915a94a4ee573aa0a7b2b017e23a0b655ca53f"} Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.433154 4677 scope.go:117] "RemoveContainer" containerID="6510391c92397ab7f26b1567ed8a627f449c9ce0cae708a46a17278bbb069e05" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.433119 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.437192 4677 generic.go:334] "Generic (PLEG): container finished" podID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerID="3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c" exitCode=0 Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.438747 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4m64" event={"ID":"e2dbbddf-0907-479f-a696-f5fecd15c4da","Type":"ContainerDied","Data":"3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c"} Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.438791 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4m64" event={"ID":"e2dbbddf-0907-479f-a696-f5fecd15c4da","Type":"ContainerStarted","Data":"9fc0f2fc43ed2adf5f7438665484d68898dfc863c6c7c60645c84c537d558c12"} Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.458767 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8322e4ce-8390-4013-a3e2-d264e5f7568f","Type":"ContainerDied","Data":"ca9db80aebf2eb7556caceea35b5201b3c76b5f76ff5fdebbe92b3c86e7c09f5"} Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.458863 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.471136 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" podStartSLOduration=3.056485013 podStartE2EDuration="13.471110142s" podCreationTimestamp="2025-12-03 14:11:51 +0000 UTC" firstStartedPulling="2025-12-03 14:11:52.818164735 +0000 UTC m=+1503.564497190" lastFinishedPulling="2025-12-03 14:12:03.232789864 +0000 UTC m=+1513.979122319" observedRunningTime="2025-12-03 14:12:04.450014429 +0000 UTC m=+1515.196346894" watchObservedRunningTime="2025-12-03 14:12:04.471110142 +0000 UTC m=+1515.217442597" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.492149 4677 scope.go:117] "RemoveContainer" containerID="76bf2f570e3bbbcf5bc48c476c42d98548b1d6d49fcb95f4d041d75dbed406f3" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.506788 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.536023 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.536792 4677 scope.go:117] "RemoveContainer" containerID="7d23fce7522889cfbe7adc1630137d136863fbc7c656995a2e903377d2616952" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.567365 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.569016 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.572544 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.573004 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.577879 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.587010 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.602115 4677 scope.go:117] "RemoveContainer" containerID="a7f778ef044e0ed9a61d6efa1f1db0f5af8e0174c2f666412f899bc8e5875a7f" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.605880 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620608 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620666 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620711 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620787 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620808 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620842 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620863 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n7gc\" (UniqueName: \"kubernetes.io/projected/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-kube-api-access-8n7gc\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.620890 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.642419 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.644343 4677 scope.go:117] "RemoveContainer" containerID="8eb03e8a7553e99a3829b7daba8b1f6493a8079ffce14c6efcbddcd626a9561f" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.647223 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.650450 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.650745 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.651937 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.677941 4677 scope.go:117] "RemoveContainer" containerID="23912dda68269bd54ba8dba59d09c5050701f64f121779ddc45dc2e6eab3bb68" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.722986 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723042 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n7gc\" (UniqueName: \"kubernetes.io/projected/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-kube-api-access-8n7gc\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723065 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723088 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-config-data\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723122 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvh4f\" (UniqueName: \"kubernetes.io/projected/c325426d-30c5-4130-842c-80be356d6ae7-kube-api-access-kvh4f\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723165 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-log-httpd\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723202 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723227 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723256 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-scripts\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723287 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723314 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723331 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723362 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723379 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.723394 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-run-httpd\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.724639 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.724686 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.725221 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.730386 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.731592 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.734393 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.741918 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n7gc\" (UniqueName: \"kubernetes.io/projected/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-kube-api-access-8n7gc\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.743993 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb0a35e-349c-4fab-ab4d-c9f1d06194cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.761269 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.765734 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: W1203 14:12:04.770273 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23eba9f1_c6e9_4dc3_ba5d_9cf53f3c7a9c.slice/crio-dd55dd9dfa083310ab56e29d27ada54558dd601123fd7815851f6065cfa45494 WatchSource:0}: Error finding container dd55dd9dfa083310ab56e29d27ada54558dd601123fd7815851f6065cfa45494: Status 404 returned error can't find the container with id dd55dd9dfa083310ab56e29d27ada54558dd601123fd7815851f6065cfa45494 Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.825771 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-run-httpd\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.825844 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-config-data\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.825883 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvh4f\" (UniqueName: \"kubernetes.io/projected/c325426d-30c5-4130-842c-80be356d6ae7-kube-api-access-kvh4f\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.825939 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-log-httpd\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.826011 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.826042 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.826081 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-scripts\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.826990 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-run-httpd\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.827765 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-log-httpd\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.834312 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-config-data\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.838174 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.838809 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.839369 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-scripts\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.852200 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.852450 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" containerID="cri-o://4308dd27ba345528b74a491ff2f517c91641232e620432c6ea4b6271efa48e9b" gracePeriod=30 Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.853834 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvh4f\" (UniqueName: \"kubernetes.io/projected/c325426d-30c5-4130-842c-80be356d6ae7-kube-api-access-kvh4f\") pod \"ceilometer-0\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " pod="openstack/ceilometer-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.903475 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:04 crc kubenswrapper[4677]: I1203 14:12:04.971879 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:05 crc kubenswrapper[4677]: I1203 14:12:05.481193 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c","Type":"ContainerStarted","Data":"dd55dd9dfa083310ab56e29d27ada54558dd601123fd7815851f6065cfa45494"} Dec 03 14:12:05 crc kubenswrapper[4677]: I1203 14:12:05.527616 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:12:05 crc kubenswrapper[4677]: I1203 14:12:05.609469 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:05 crc kubenswrapper[4677]: I1203 14:12:05.994578 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232a4115-1312-485b-8df6-cad4097e7b6f" path="/var/lib/kubelet/pods/232a4115-1312-485b-8df6-cad4097e7b6f/volumes" Dec 03 14:12:05 crc kubenswrapper[4677]: I1203 14:12:05.995572 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8322e4ce-8390-4013-a3e2-d264e5f7568f" path="/var/lib/kubelet/pods/8322e4ce-8390-4013-a3e2-d264e5f7568f/volumes" Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.502403 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4m64" event={"ID":"e2dbbddf-0907-479f-a696-f5fecd15c4da","Type":"ContainerStarted","Data":"814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.515433 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerStarted","Data":"ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.515473 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerStarted","Data":"ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.515486 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerStarted","Data":"ca8655f3c74248a0994cd2a3992518c29b9f5d0b5f1d79645810ff8955ef158f"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.520468 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc","Type":"ContainerStarted","Data":"56efc9d84cd68b18bbe6c9ff32123f1dfb15e55daa4cbafd8c7dd8bc1306eeca"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.520526 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc","Type":"ContainerStarted","Data":"cf4d66b271ca3d4982b61f6a8ce356c7fc8e80b24e370ecaeb8fa8162b6ffd67"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.527845 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c","Type":"ContainerStarted","Data":"e2da791dd4b1ed6d9714eda235d205f97adc7e346b77d6a142a143734b5d8541"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.527891 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c","Type":"ContainerStarted","Data":"b31b4525d7728d690fe6720277eb442c59b07808fdd788d897dd5ad0bad54a6c"} Dec 03 14:12:06 crc kubenswrapper[4677]: I1203 14:12:06.556628 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.55660125 podStartE2EDuration="3.55660125s" podCreationTimestamp="2025-12-03 14:12:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:06.548636322 +0000 UTC m=+1517.294968797" watchObservedRunningTime="2025-12-03 14:12:06.55660125 +0000 UTC m=+1517.302933725" Dec 03 14:12:07 crc kubenswrapper[4677]: I1203 14:12:07.294257 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:07 crc kubenswrapper[4677]: I1203 14:12:07.542989 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerStarted","Data":"8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b"} Dec 03 14:12:07 crc kubenswrapper[4677]: I1203 14:12:07.545554 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8eb0a35e-349c-4fab-ab4d-c9f1d06194cc","Type":"ContainerStarted","Data":"db9dc70cd57be8d078a001a5e5d4fe4aa8f573e7393bd10c04fa93a859dd6343"} Dec 03 14:12:08 crc kubenswrapper[4677]: I1203 14:12:08.594843 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.594820623 podStartE2EDuration="4.594820623s" podCreationTimestamp="2025-12-03 14:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:08.581233197 +0000 UTC m=+1519.327565662" watchObservedRunningTime="2025-12-03 14:12:08.594820623 +0000 UTC m=+1519.341153088" Dec 03 14:12:09 crc kubenswrapper[4677]: I1203 14:12:09.574724 4677 generic.go:334] "Generic (PLEG): container finished" podID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerID="814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9" exitCode=0 Dec 03 14:12:09 crc kubenswrapper[4677]: I1203 14:12:09.574764 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4m64" event={"ID":"e2dbbddf-0907-479f-a696-f5fecd15c4da","Type":"ContainerDied","Data":"814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9"} Dec 03 14:12:10 crc kubenswrapper[4677]: I1203 14:12:10.597458 4677 generic.go:334] "Generic (PLEG): container finished" podID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerID="4308dd27ba345528b74a491ff2f517c91641232e620432c6ea4b6271efa48e9b" exitCode=0 Dec 03 14:12:10 crc kubenswrapper[4677]: I1203 14:12:10.597534 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerDied","Data":"4308dd27ba345528b74a491ff2f517c91641232e620432c6ea4b6271efa48e9b"} Dec 03 14:12:10 crc kubenswrapper[4677]: I1203 14:12:10.597585 4677 scope.go:117] "RemoveContainer" containerID="bb94a31a78ea7f1f890f5acfc3d4025633dc5031a4146928590a9ab6eda17b77" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.116715 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.216209 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f93ff7e-f6ab-4c00-8284-70a4354e576a-logs\") pod \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.216679 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-combined-ca-bundle\") pod \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.216715 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cfjh\" (UniqueName: \"kubernetes.io/projected/0f93ff7e-f6ab-4c00-8284-70a4354e576a-kube-api-access-9cfjh\") pod \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.216708 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f93ff7e-f6ab-4c00-8284-70a4354e576a-logs" (OuterVolumeSpecName: "logs") pod "0f93ff7e-f6ab-4c00-8284-70a4354e576a" (UID: "0f93ff7e-f6ab-4c00-8284-70a4354e576a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.216804 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-custom-prometheus-ca\") pod \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.216826 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-config-data\") pod \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\" (UID: \"0f93ff7e-f6ab-4c00-8284-70a4354e576a\") " Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.217215 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f93ff7e-f6ab-4c00-8284-70a4354e576a-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.224809 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f93ff7e-f6ab-4c00-8284-70a4354e576a-kube-api-access-9cfjh" (OuterVolumeSpecName: "kube-api-access-9cfjh") pod "0f93ff7e-f6ab-4c00-8284-70a4354e576a" (UID: "0f93ff7e-f6ab-4c00-8284-70a4354e576a"). InnerVolumeSpecName "kube-api-access-9cfjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.249697 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f93ff7e-f6ab-4c00-8284-70a4354e576a" (UID: "0f93ff7e-f6ab-4c00-8284-70a4354e576a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.251523 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "0f93ff7e-f6ab-4c00-8284-70a4354e576a" (UID: "0f93ff7e-f6ab-4c00-8284-70a4354e576a"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.273401 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-config-data" (OuterVolumeSpecName: "config-data") pod "0f93ff7e-f6ab-4c00-8284-70a4354e576a" (UID: "0f93ff7e-f6ab-4c00-8284-70a4354e576a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.319203 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.319252 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cfjh\" (UniqueName: \"kubernetes.io/projected/0f93ff7e-f6ab-4c00-8284-70a4354e576a-kube-api-access-9cfjh\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.319274 4677 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.319296 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f93ff7e-f6ab-4c00-8284-70a4354e576a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:12 crc kubenswrapper[4677]: E1203 14:12:12.531341 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice/crio-5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice/crio-1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991\": RecentStats: unable to find data in memory cache]" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.623420 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4m64" event={"ID":"e2dbbddf-0907-479f-a696-f5fecd15c4da","Type":"ContainerStarted","Data":"c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867"} Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.627590 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerStarted","Data":"dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c"} Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.627764 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-central-agent" containerID="cri-o://ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d" gracePeriod=30 Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.628007 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.628056 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="proxy-httpd" containerID="cri-o://dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c" gracePeriod=30 Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.628113 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="sg-core" containerID="cri-o://8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b" gracePeriod=30 Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.628157 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-notification-agent" containerID="cri-o://ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f" gracePeriod=30 Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.631074 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f93ff7e-f6ab-4c00-8284-70a4354e576a","Type":"ContainerDied","Data":"52ff4c1ad22782bc249cca16979fb7c08be743778a30945ce2c47217347856ff"} Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.631117 4677 scope.go:117] "RemoveContainer" containerID="4308dd27ba345528b74a491ff2f517c91641232e620432c6ea4b6271efa48e9b" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.631254 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.670682 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x4m64" podStartSLOduration=7.235790764 podStartE2EDuration="14.670647741s" podCreationTimestamp="2025-12-03 14:11:58 +0000 UTC" firstStartedPulling="2025-12-03 14:12:04.44163552 +0000 UTC m=+1515.187967975" lastFinishedPulling="2025-12-03 14:12:11.876492497 +0000 UTC m=+1522.622824952" observedRunningTime="2025-12-03 14:12:12.64616626 +0000 UTC m=+1523.392498735" watchObservedRunningTime="2025-12-03 14:12:12.670647741 +0000 UTC m=+1523.416980196" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.682476 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.536400242 podStartE2EDuration="8.68245978s" podCreationTimestamp="2025-12-03 14:12:04 +0000 UTC" firstStartedPulling="2025-12-03 14:12:05.627783393 +0000 UTC m=+1516.374115848" lastFinishedPulling="2025-12-03 14:12:11.773842891 +0000 UTC m=+1522.520175386" observedRunningTime="2025-12-03 14:12:12.668439593 +0000 UTC m=+1523.414772048" watchObservedRunningTime="2025-12-03 14:12:12.68245978 +0000 UTC m=+1523.428792245" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.693868 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.703069 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.714854 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:12:12 crc kubenswrapper[4677]: E1203 14:12:12.715434 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.715450 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: E1203 14:12:12.715469 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.715477 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: E1203 14:12:12.715494 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.715500 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: E1203 14:12:12.715513 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.715519 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.715697 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.715709 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.715725 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.716381 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.723478 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.725935 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.726067 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.726145 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-logs\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.726282 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcwpp\" (UniqueName: \"kubernetes.io/projected/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-kube-api-access-wcwpp\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.726349 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.730595 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.827115 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.827194 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.827253 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-logs\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.827342 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcwpp\" (UniqueName: \"kubernetes.io/projected/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-kube-api-access-wcwpp\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.827395 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.827686 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-logs\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.832902 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.832908 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.833693 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:12 crc kubenswrapper[4677]: I1203 14:12:12.846889 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcwpp\" (UniqueName: \"kubernetes.io/projected/8ed2ece8-e640-495d-b5fb-b0cd73a190e2-kube-api-access-wcwpp\") pod \"watcher-decision-engine-0\" (UID: \"8ed2ece8-e640-495d-b5fb-b0cd73a190e2\") " pod="openstack/watcher-decision-engine-0" Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.066018 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.573864 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 14:12:13 crc kubenswrapper[4677]: W1203 14:12:13.577075 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ed2ece8_e640_495d_b5fb_b0cd73a190e2.slice/crio-765062c56cf3feb7d530bb3a2d6c74ed2bb5e61d12935055308a8ff57beaada0 WatchSource:0}: Error finding container 765062c56cf3feb7d530bb3a2d6c74ed2bb5e61d12935055308a8ff57beaada0: Status 404 returned error can't find the container with id 765062c56cf3feb7d530bb3a2d6c74ed2bb5e61d12935055308a8ff57beaada0 Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.667034 4677 generic.go:334] "Generic (PLEG): container finished" podID="c325426d-30c5-4130-842c-80be356d6ae7" containerID="dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c" exitCode=0 Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.667134 4677 generic.go:334] "Generic (PLEG): container finished" podID="c325426d-30c5-4130-842c-80be356d6ae7" containerID="8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b" exitCode=2 Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.667145 4677 generic.go:334] "Generic (PLEG): container finished" podID="c325426d-30c5-4130-842c-80be356d6ae7" containerID="ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f" exitCode=0 Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.667192 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerDied","Data":"dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c"} Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.667225 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerDied","Data":"8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b"} Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.667237 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerDied","Data":"ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f"} Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.669263 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ed2ece8-e640-495d-b5fb-b0cd73a190e2","Type":"ContainerStarted","Data":"765062c56cf3feb7d530bb3a2d6c74ed2bb5e61d12935055308a8ff57beaada0"} Dec 03 14:12:13 crc kubenswrapper[4677]: I1203 14:12:13.991535 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" path="/var/lib/kubelet/pods/0f93ff7e-f6ab-4c00-8284-70a4354e576a/volumes" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.222517 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.222570 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.264738 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.285532 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.682961 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8ed2ece8-e640-495d-b5fb-b0cd73a190e2","Type":"ContainerStarted","Data":"2a11c915d6f94c15fdc106ea5eebe043330de6376e7403973df6b1e92f7396f2"} Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.685747 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.685814 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.712081 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.712060846 podStartE2EDuration="2.712060846s" podCreationTimestamp="2025-12-03 14:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:14.709696805 +0000 UTC m=+1525.456029290" watchObservedRunningTime="2025-12-03 14:12:14.712060846 +0000 UTC m=+1525.458393331" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.948413 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.948474 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:14 crc kubenswrapper[4677]: I1203 14:12:14.977798 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:15 crc kubenswrapper[4677]: I1203 14:12:15.002103 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:15 crc kubenswrapper[4677]: I1203 14:12:15.692268 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:15 crc kubenswrapper[4677]: I1203 14:12:15.692744 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:16 crc kubenswrapper[4677]: I1203 14:12:16.827035 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:12:16 crc kubenswrapper[4677]: I1203 14:12:16.828713 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:12:17 crc kubenswrapper[4677]: I1203 14:12:17.925344 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:17 crc kubenswrapper[4677]: I1203 14:12:17.925705 4677 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:12:18 crc kubenswrapper[4677]: I1203 14:12:18.073471 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:12:18 crc kubenswrapper[4677]: I1203 14:12:18.764900 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:12:18 crc kubenswrapper[4677]: I1203 14:12:18.764985 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.579082 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.692337 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-config-data\") pod \"c325426d-30c5-4130-842c-80be356d6ae7\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.692752 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-run-httpd\") pod \"c325426d-30c5-4130-842c-80be356d6ae7\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.692809 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-combined-ca-bundle\") pod \"c325426d-30c5-4130-842c-80be356d6ae7\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.692835 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-sg-core-conf-yaml\") pod \"c325426d-30c5-4130-842c-80be356d6ae7\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.692896 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-scripts\") pod \"c325426d-30c5-4130-842c-80be356d6ae7\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.693141 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvh4f\" (UniqueName: \"kubernetes.io/projected/c325426d-30c5-4130-842c-80be356d6ae7-kube-api-access-kvh4f\") pod \"c325426d-30c5-4130-842c-80be356d6ae7\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.693206 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-log-httpd\") pod \"c325426d-30c5-4130-842c-80be356d6ae7\" (UID: \"c325426d-30c5-4130-842c-80be356d6ae7\") " Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.693639 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c325426d-30c5-4130-842c-80be356d6ae7" (UID: "c325426d-30c5-4130-842c-80be356d6ae7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.693875 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c325426d-30c5-4130-842c-80be356d6ae7" (UID: "c325426d-30c5-4130-842c-80be356d6ae7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.693912 4677 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.699471 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c325426d-30c5-4130-842c-80be356d6ae7-kube-api-access-kvh4f" (OuterVolumeSpecName: "kube-api-access-kvh4f") pod "c325426d-30c5-4130-842c-80be356d6ae7" (UID: "c325426d-30c5-4130-842c-80be356d6ae7"). InnerVolumeSpecName "kube-api-access-kvh4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.700055 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-scripts" (OuterVolumeSpecName: "scripts") pod "c325426d-30c5-4130-842c-80be356d6ae7" (UID: "c325426d-30c5-4130-842c-80be356d6ae7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.753564 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c325426d-30c5-4130-842c-80be356d6ae7" (UID: "c325426d-30c5-4130-842c-80be356d6ae7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.778185 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c325426d-30c5-4130-842c-80be356d6ae7" (UID: "c325426d-30c5-4130-842c-80be356d6ae7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.796298 4677 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c325426d-30c5-4130-842c-80be356d6ae7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.796361 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.796376 4677 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.796390 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.796401 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvh4f\" (UniqueName: \"kubernetes.io/projected/c325426d-30c5-4130-842c-80be356d6ae7-kube-api-access-kvh4f\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.813232 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x4m64" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="registry-server" probeResult="failure" output=< Dec 03 14:12:19 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 14:12:19 crc kubenswrapper[4677]: > Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.818861 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-config-data" (OuterVolumeSpecName: "config-data") pod "c325426d-30c5-4130-842c-80be356d6ae7" (UID: "c325426d-30c5-4130-842c-80be356d6ae7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.898218 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c325426d-30c5-4130-842c-80be356d6ae7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.915237 4677 generic.go:334] "Generic (PLEG): container finished" podID="c325426d-30c5-4130-842c-80be356d6ae7" containerID="ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d" exitCode=0 Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.915292 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerDied","Data":"ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d"} Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.915326 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c325426d-30c5-4130-842c-80be356d6ae7","Type":"ContainerDied","Data":"ca8655f3c74248a0994cd2a3992518c29b9f5d0b5f1d79645810ff8955ef158f"} Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.915348 4677 scope.go:117] "RemoveContainer" containerID="dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.915348 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.960398 4677 scope.go:117] "RemoveContainer" containerID="8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b" Dec 03 14:12:19 crc kubenswrapper[4677]: I1203 14:12:19.993772 4677 scope.go:117] "RemoveContainer" containerID="ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.007384 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.008267 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.024823 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.025237 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="sg-core" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025255 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="sg-core" Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.025268 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-central-agent" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025274 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-central-agent" Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.025292 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-notification-agent" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025298 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-notification-agent" Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.025317 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="proxy-httpd" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025323 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="proxy-httpd" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025520 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="proxy-httpd" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025540 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="sg-core" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025551 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-central-agent" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025560 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c325426d-30c5-4130-842c-80be356d6ae7" containerName="ceilometer-notification-agent" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.025567 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f93ff7e-f6ab-4c00-8284-70a4354e576a" containerName="watcher-decision-engine" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.027381 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.031300 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.031457 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.036398 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.036571 4677 scope.go:117] "RemoveContainer" containerID="ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.075834 4677 scope.go:117] "RemoveContainer" containerID="dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c" Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.076258 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c\": container with ID starting with dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c not found: ID does not exist" containerID="dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.076304 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c"} err="failed to get container status \"dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c\": rpc error: code = NotFound desc = could not find container \"dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c\": container with ID starting with dc01e5ede7ee06f9790cdcf50c3add458ad0bd3eb0d484b37cec2112324bdc1c not found: ID does not exist" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.076330 4677 scope.go:117] "RemoveContainer" containerID="8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b" Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.076623 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b\": container with ID starting with 8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b not found: ID does not exist" containerID="8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.076647 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b"} err="failed to get container status \"8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b\": rpc error: code = NotFound desc = could not find container \"8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b\": container with ID starting with 8cfd6cd378a804297160054607c05d774e3121e8e4ceb94e15063415f4c8e81b not found: ID does not exist" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.076678 4677 scope.go:117] "RemoveContainer" containerID="ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f" Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.076892 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f\": container with ID starting with ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f not found: ID does not exist" containerID="ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.076914 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f"} err="failed to get container status \"ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f\": rpc error: code = NotFound desc = could not find container \"ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f\": container with ID starting with ff546cabf162a4f78b76b0f0e91dd1751fcf5c95670e86970e8052caa9cfa37f not found: ID does not exist" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.076926 4677 scope.go:117] "RemoveContainer" containerID="ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d" Dec 03 14:12:20 crc kubenswrapper[4677]: E1203 14:12:20.077204 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d\": container with ID starting with ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d not found: ID does not exist" containerID="ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.077237 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d"} err="failed to get container status \"ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d\": rpc error: code = NotFound desc = could not find container \"ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d\": container with ID starting with ad3ebb51ca62401c47eee5b5344d0c4b907d517b6fada1152db6146daf01ad7d not found: ID does not exist" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.102726 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-run-httpd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.102797 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-log-httpd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.102842 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.102875 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-scripts\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.102977 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.103118 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-config-data\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.103175 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkjmd\" (UniqueName: \"kubernetes.io/projected/c0b61325-6e17-4a89-a83a-0f1de7d2db43-kube-api-access-jkjmd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.204703 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkjmd\" (UniqueName: \"kubernetes.io/projected/c0b61325-6e17-4a89-a83a-0f1de7d2db43-kube-api-access-jkjmd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.204830 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-run-httpd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.204873 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-log-httpd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.205317 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.205369 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-scripts\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.205414 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.205433 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-run-httpd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.205508 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-config-data\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.206088 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-log-httpd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.210880 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-scripts\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.212627 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-config-data\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.211649 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.214485 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.222720 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkjmd\" (UniqueName: \"kubernetes.io/projected/c0b61325-6e17-4a89-a83a-0f1de7d2db43-kube-api-access-jkjmd\") pod \"ceilometer-0\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.366060 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:12:20 crc kubenswrapper[4677]: W1203 14:12:20.829126 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0b61325_6e17_4a89_a83a_0f1de7d2db43.slice/crio-259e232ac64fa65069d8f9ea390ebf0f47bbd9bb43bd76cf9df41539366601e4 WatchSource:0}: Error finding container 259e232ac64fa65069d8f9ea390ebf0f47bbd9bb43bd76cf9df41539366601e4: Status 404 returned error can't find the container with id 259e232ac64fa65069d8f9ea390ebf0f47bbd9bb43bd76cf9df41539366601e4 Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.830008 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:12:20 crc kubenswrapper[4677]: I1203 14:12:20.924720 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerStarted","Data":"259e232ac64fa65069d8f9ea390ebf0f47bbd9bb43bd76cf9df41539366601e4"} Dec 03 14:12:21 crc kubenswrapper[4677]: I1203 14:12:21.934828 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerStarted","Data":"c30f57bfb3109e0309985c74f4a034508e57c525ec9a66c240862986887d6ded"} Dec 03 14:12:21 crc kubenswrapper[4677]: I1203 14:12:21.935020 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerStarted","Data":"b91a5deb92cad98bcb2c8b1bb662cde4d194cf37b5175dc43b5133efd8c2a20c"} Dec 03 14:12:21 crc kubenswrapper[4677]: I1203 14:12:21.987319 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c325426d-30c5-4130-842c-80be356d6ae7" path="/var/lib/kubelet/pods/c325426d-30c5-4130-842c-80be356d6ae7/volumes" Dec 03 14:12:22 crc kubenswrapper[4677]: E1203 14:12:22.796293 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice/crio-5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice/crio-1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:12:22 crc kubenswrapper[4677]: I1203 14:12:22.946230 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerStarted","Data":"5331a97bb1a0493c45473f17b93463b4d3a106cc4efdaff51a3e290f7c2585be"} Dec 03 14:12:22 crc kubenswrapper[4677]: I1203 14:12:22.948844 4677 generic.go:334] "Generic (PLEG): container finished" podID="b9b349d0-d7aa-45cb-aab4-d50379b27c4d" containerID="047d9dc6493b6bc1a221b6c2bdad811a81529f68519c088a20b717a7fd6bbd15" exitCode=0 Dec 03 14:12:22 crc kubenswrapper[4677]: I1203 14:12:22.948872 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" event={"ID":"b9b349d0-d7aa-45cb-aab4-d50379b27c4d","Type":"ContainerDied","Data":"047d9dc6493b6bc1a221b6c2bdad811a81529f68519c088a20b717a7fd6bbd15"} Dec 03 14:12:23 crc kubenswrapper[4677]: I1203 14:12:23.067708 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:23 crc kubenswrapper[4677]: I1203 14:12:23.098983 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:23 crc kubenswrapper[4677]: I1203 14:12:23.958386 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.000017 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.326133 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.381517 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-combined-ca-bundle\") pod \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.381605 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtdg5\" (UniqueName: \"kubernetes.io/projected/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-kube-api-access-wtdg5\") pod \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.381772 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-config-data\") pod \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.383100 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-scripts\") pod \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\" (UID: \"b9b349d0-d7aa-45cb-aab4-d50379b27c4d\") " Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.414654 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-scripts" (OuterVolumeSpecName: "scripts") pod "b9b349d0-d7aa-45cb-aab4-d50379b27c4d" (UID: "b9b349d0-d7aa-45cb-aab4-d50379b27c4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.417116 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-kube-api-access-wtdg5" (OuterVolumeSpecName: "kube-api-access-wtdg5") pod "b9b349d0-d7aa-45cb-aab4-d50379b27c4d" (UID: "b9b349d0-d7aa-45cb-aab4-d50379b27c4d"). InnerVolumeSpecName "kube-api-access-wtdg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.433769 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-config-data" (OuterVolumeSpecName: "config-data") pod "b9b349d0-d7aa-45cb-aab4-d50379b27c4d" (UID: "b9b349d0-d7aa-45cb-aab4-d50379b27c4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.439240 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9b349d0-d7aa-45cb-aab4-d50379b27c4d" (UID: "b9b349d0-d7aa-45cb-aab4-d50379b27c4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.485531 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.485853 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtdg5\" (UniqueName: \"kubernetes.io/projected/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-kube-api-access-wtdg5\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.485865 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.485875 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9b349d0-d7aa-45cb-aab4-d50379b27c4d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.969322 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.969329 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkxw5" event={"ID":"b9b349d0-d7aa-45cb-aab4-d50379b27c4d","Type":"ContainerDied","Data":"3341af9480ba9a4a35e454f1f41316beb7b71c42474f16bf12c4d3ee22eac172"} Dec 03 14:12:24 crc kubenswrapper[4677]: I1203 14:12:24.969507 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3341af9480ba9a4a35e454f1f41316beb7b71c42474f16bf12c4d3ee22eac172" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.088422 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 14:12:25 crc kubenswrapper[4677]: E1203 14:12:25.089061 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b349d0-d7aa-45cb-aab4-d50379b27c4d" containerName="nova-cell0-conductor-db-sync" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.089092 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b349d0-d7aa-45cb-aab4-d50379b27c4d" containerName="nova-cell0-conductor-db-sync" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.089364 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b349d0-d7aa-45cb-aab4-d50379b27c4d" containerName="nova-cell0-conductor-db-sync" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.090285 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.093008 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.097563 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s7d94" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.108098 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.203791 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65d357-06ac-49f7-85d0-890bba2de21d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.203918 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t4t9\" (UniqueName: \"kubernetes.io/projected/fd65d357-06ac-49f7-85d0-890bba2de21d-kube-api-access-8t4t9\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.204052 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65d357-06ac-49f7-85d0-890bba2de21d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.305918 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t4t9\" (UniqueName: \"kubernetes.io/projected/fd65d357-06ac-49f7-85d0-890bba2de21d-kube-api-access-8t4t9\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.306069 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65d357-06ac-49f7-85d0-890bba2de21d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.306111 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65d357-06ac-49f7-85d0-890bba2de21d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.316821 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd65d357-06ac-49f7-85d0-890bba2de21d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.318676 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd65d357-06ac-49f7-85d0-890bba2de21d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.337565 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t4t9\" (UniqueName: \"kubernetes.io/projected/fd65d357-06ac-49f7-85d0-890bba2de21d-kube-api-access-8t4t9\") pod \"nova-cell0-conductor-0\" (UID: \"fd65d357-06ac-49f7-85d0-890bba2de21d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.416382 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.897235 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 14:12:25 crc kubenswrapper[4677]: I1203 14:12:25.988932 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fd65d357-06ac-49f7-85d0-890bba2de21d","Type":"ContainerStarted","Data":"f4204b76d39a01202b7f1ddc89a87b48b8261ef6ac15891f68bf7d8fbea299e4"} Dec 03 14:12:26 crc kubenswrapper[4677]: I1203 14:12:26.993216 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fd65d357-06ac-49f7-85d0-890bba2de21d","Type":"ContainerStarted","Data":"49e00d430115b792e43559a6c62939ca7f7ff1d7a78e29befaef9cfa49cb5b6c"} Dec 03 14:12:26 crc kubenswrapper[4677]: I1203 14:12:26.993489 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:27 crc kubenswrapper[4677]: I1203 14:12:27.013168 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.013139487 podStartE2EDuration="2.013139487s" podCreationTimestamp="2025-12-03 14:12:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:27.011170545 +0000 UTC m=+1537.757503040" watchObservedRunningTime="2025-12-03 14:12:27.013139487 +0000 UTC m=+1537.759471972" Dec 03 14:12:28 crc kubenswrapper[4677]: I1203 14:12:28.817697 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:12:28 crc kubenswrapper[4677]: I1203 14:12:28.867091 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:12:29 crc kubenswrapper[4677]: I1203 14:12:29.638734 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x4m64"] Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.027680 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x4m64" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="registry-server" containerID="cri-o://c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867" gracePeriod=2 Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.478102 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.513699 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-utilities\") pod \"e2dbbddf-0907-479f-a696-f5fecd15c4da\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.513758 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhgmk\" (UniqueName: \"kubernetes.io/projected/e2dbbddf-0907-479f-a696-f5fecd15c4da-kube-api-access-vhgmk\") pod \"e2dbbddf-0907-479f-a696-f5fecd15c4da\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.514206 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-catalog-content\") pod \"e2dbbddf-0907-479f-a696-f5fecd15c4da\" (UID: \"e2dbbddf-0907-479f-a696-f5fecd15c4da\") " Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.515967 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-utilities" (OuterVolumeSpecName: "utilities") pod "e2dbbddf-0907-479f-a696-f5fecd15c4da" (UID: "e2dbbddf-0907-479f-a696-f5fecd15c4da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.519890 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2dbbddf-0907-479f-a696-f5fecd15c4da-kube-api-access-vhgmk" (OuterVolumeSpecName: "kube-api-access-vhgmk") pod "e2dbbddf-0907-479f-a696-f5fecd15c4da" (UID: "e2dbbddf-0907-479f-a696-f5fecd15c4da"). InnerVolumeSpecName "kube-api-access-vhgmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.608595 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2dbbddf-0907-479f-a696-f5fecd15c4da" (UID: "e2dbbddf-0907-479f-a696-f5fecd15c4da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.616077 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.616755 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhgmk\" (UniqueName: \"kubernetes.io/projected/e2dbbddf-0907-479f-a696-f5fecd15c4da-kube-api-access-vhgmk\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:30 crc kubenswrapper[4677]: I1203 14:12:30.616877 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2dbbddf-0907-479f-a696-f5fecd15c4da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.042210 4677 generic.go:334] "Generic (PLEG): container finished" podID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerID="c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867" exitCode=0 Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.042294 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4m64" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.042323 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4m64" event={"ID":"e2dbbddf-0907-479f-a696-f5fecd15c4da","Type":"ContainerDied","Data":"c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867"} Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.043839 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4m64" event={"ID":"e2dbbddf-0907-479f-a696-f5fecd15c4da","Type":"ContainerDied","Data":"9fc0f2fc43ed2adf5f7438665484d68898dfc863c6c7c60645c84c537d558c12"} Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.043907 4677 scope.go:117] "RemoveContainer" containerID="c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.082324 4677 scope.go:117] "RemoveContainer" containerID="814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.088337 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x4m64"] Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.100455 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x4m64"] Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.115542 4677 scope.go:117] "RemoveContainer" containerID="3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.149470 4677 scope.go:117] "RemoveContainer" containerID="c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867" Dec 03 14:12:31 crc kubenswrapper[4677]: E1203 14:12:31.149885 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867\": container with ID starting with c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867 not found: ID does not exist" containerID="c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.149959 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867"} err="failed to get container status \"c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867\": rpc error: code = NotFound desc = could not find container \"c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867\": container with ID starting with c5fd70353026130bcbfe3e03f6d516624eba4d24fb210f3e90e00f8dccaf2867 not found: ID does not exist" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.149980 4677 scope.go:117] "RemoveContainer" containerID="814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9" Dec 03 14:12:31 crc kubenswrapper[4677]: E1203 14:12:31.150257 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9\": container with ID starting with 814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9 not found: ID does not exist" containerID="814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.150284 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9"} err="failed to get container status \"814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9\": rpc error: code = NotFound desc = could not find container \"814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9\": container with ID starting with 814b05eed369ab272694bc9b43a905ecd090b1e60b741cd2b12b0a77937023e9 not found: ID does not exist" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.150305 4677 scope.go:117] "RemoveContainer" containerID="3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c" Dec 03 14:12:31 crc kubenswrapper[4677]: E1203 14:12:31.150516 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c\": container with ID starting with 3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c not found: ID does not exist" containerID="3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.150538 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c"} err="failed to get container status \"3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c\": rpc error: code = NotFound desc = could not find container \"3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c\": container with ID starting with 3d47ecf7a474a333b248e46de86fb7a2fc588354daaf53901df866719b66e10c not found: ID does not exist" Dec 03 14:12:31 crc kubenswrapper[4677]: I1203 14:12:31.989679 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" path="/var/lib/kubelet/pods/e2dbbddf-0907-479f-a696-f5fecd15c4da/volumes" Dec 03 14:12:33 crc kubenswrapper[4677]: E1203 14:12:33.043574 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice/crio-5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice/crio-1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:12:33 crc kubenswrapper[4677]: E1203 14:12:33.771740 4677 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: Get \"https://access.redhat.com/webassets/docker/content/sigstore/ubi9/httpd-24@sha256=379ba2fc60106c2026bafa8e7d0ad585d87e6e7fe3f0a6a3d714f744e450bdf2/signature-1\": net/http: TLS handshake timeout" image="registry.redhat.io/ubi9/httpd-24:latest" Dec 03 14:12:33 crc kubenswrapper[4677]: E1203 14:12:33.771936 4677 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jkjmd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(c0b61325-6e17-4a89-a83a-0f1de7d2db43): ErrImagePull: copying system image from manifest list: reading signatures: Get \"https://access.redhat.com/webassets/docker/content/sigstore/ubi9/httpd-24@sha256=379ba2fc60106c2026bafa8e7d0ad585d87e6e7fe3f0a6a3d714f744e450bdf2/signature-1\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 14:12:33 crc kubenswrapper[4677]: E1203 14:12:33.773216 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"copying system image from manifest list: reading signatures: Get \\\"https://access.redhat.com/webassets/docker/content/sigstore/ubi9/httpd-24@sha256=379ba2fc60106c2026bafa8e7d0ad585d87e6e7fe3f0a6a3d714f744e450bdf2/signature-1\\\": net/http: TLS handshake timeout\"" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" Dec 03 14:12:34 crc kubenswrapper[4677]: E1203 14:12:34.071632 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/ubi9/httpd-24:latest\\\"\"" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.451752 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.973857 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-kjshq"] Dec 03 14:12:35 crc kubenswrapper[4677]: E1203 14:12:35.976887 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="extract-utilities" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.976918 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="extract-utilities" Dec 03 14:12:35 crc kubenswrapper[4677]: E1203 14:12:35.976940 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="extract-content" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.976969 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="extract-content" Dec 03 14:12:35 crc kubenswrapper[4677]: E1203 14:12:35.977007 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="registry-server" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.977018 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="registry-server" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.977254 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2dbbddf-0907-479f-a696-f5fecd15c4da" containerName="registry-server" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.978307 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.984482 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 14:12:35 crc kubenswrapper[4677]: I1203 14:12:35.984613 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.002571 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kjshq"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.026930 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nljkw\" (UniqueName: \"kubernetes.io/projected/fed8b246-9742-4b78-a630-b820a90ec8c8-kube-api-access-nljkw\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.027764 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.027936 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-config-data\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.029549 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-scripts\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.132340 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-config-data\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.132440 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-scripts\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.132507 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nljkw\" (UniqueName: \"kubernetes.io/projected/fed8b246-9742-4b78-a630-b820a90ec8c8-kube-api-access-nljkw\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.132609 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.143365 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-scripts\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.143571 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.143914 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-config-data\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.179251 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.180458 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.191476 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nljkw\" (UniqueName: \"kubernetes.io/projected/fed8b246-9742-4b78-a630-b820a90ec8c8-kube-api-access-nljkw\") pod \"nova-cell0-cell-mapping-kjshq\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.194038 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.211254 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.241978 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.242064 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z72hj\" (UniqueName: \"kubernetes.io/projected/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-kube-api-access-z72hj\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.242469 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.309563 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.314361 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.330635 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.333554 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.345368 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.345438 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z72hj\" (UniqueName: \"kubernetes.io/projected/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-kube-api-access-z72hj\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.345545 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.353445 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.386663 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.388451 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.394191 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z72hj\" (UniqueName: \"kubernetes.io/projected/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-kube-api-access-z72hj\") pod \"nova-cell1-novncproxy-0\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.422748 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.424052 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.427529 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.462868 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.463127 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.463161 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-config-data\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.463288 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-config-data\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.463374 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skrf2\" (UniqueName: \"kubernetes.io/projected/37e0ed04-8f63-4fb6-9598-a4199b4b7258-kube-api-access-skrf2\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.463424 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rxg8\" (UniqueName: \"kubernetes.io/projected/7c94b290-b422-470d-bf5d-f8dd6d28364d-kube-api-access-7rxg8\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.463503 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c94b290-b422-470d-bf5d-f8dd6d28364d-logs\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.482309 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.525335 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86755f6c8c-s8sbl"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.527268 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.539363 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86755f6c8c-s8sbl"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.572527 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-config-data\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.572674 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skrf2\" (UniqueName: \"kubernetes.io/projected/37e0ed04-8f63-4fb6-9598-a4199b4b7258-kube-api-access-skrf2\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.572744 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rxg8\" (UniqueName: \"kubernetes.io/projected/7c94b290-b422-470d-bf5d-f8dd6d28364d-kube-api-access-7rxg8\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.572868 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-swift-storage-0\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.572896 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c94b290-b422-470d-bf5d-f8dd6d28364d-logs\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.572938 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-nb\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.573040 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-config\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.573099 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.573162 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-svc\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.573324 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-sb\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.573452 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.573496 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-config-data\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.573641 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5h8g\" (UniqueName: \"kubernetes.io/projected/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-kube-api-access-r5h8g\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.574609 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.576094 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c94b290-b422-470d-bf5d-f8dd6d28364d-logs\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.578989 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.581400 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-config-data\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.581417 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.582472 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.584650 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.593516 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.600437 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.603856 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-config-data\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.609608 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skrf2\" (UniqueName: \"kubernetes.io/projected/37e0ed04-8f63-4fb6-9598-a4199b4b7258-kube-api-access-skrf2\") pod \"nova-scheduler-0\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.619062 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rxg8\" (UniqueName: \"kubernetes.io/projected/7c94b290-b422-470d-bf5d-f8dd6d28364d-kube-api-access-7rxg8\") pod \"nova-metadata-0\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.639284 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.683365 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-nb\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.686358 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-nb\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.693196 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-config\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.693326 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-svc\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.693472 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-sb\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.693727 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5h8g\" (UniqueName: \"kubernetes.io/projected/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-kube-api-access-r5h8g\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.693812 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.693901 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzmrt\" (UniqueName: \"kubernetes.io/projected/58617ccc-ce55-4b21-816e-6a436b52b1a2-kube-api-access-mzmrt\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.693918 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-config-data\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.694036 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58617ccc-ce55-4b21-816e-6a436b52b1a2-logs\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.694115 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-swift-storage-0\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.695598 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-config\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.696287 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-swift-storage-0\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.696656 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-sb\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.697621 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-svc\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.712212 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5h8g\" (UniqueName: \"kubernetes.io/projected/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-kube-api-access-r5h8g\") pod \"dnsmasq-dns-86755f6c8c-s8sbl\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.796458 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.796529 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzmrt\" (UniqueName: \"kubernetes.io/projected/58617ccc-ce55-4b21-816e-6a436b52b1a2-kube-api-access-mzmrt\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.796560 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-config-data\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.796608 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58617ccc-ce55-4b21-816e-6a436b52b1a2-logs\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.797469 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58617ccc-ce55-4b21-816e-6a436b52b1a2-logs\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.814714 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-config-data\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.815132 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.834733 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzmrt\" (UniqueName: \"kubernetes.io/projected/58617ccc-ce55-4b21-816e-6a436b52b1a2-kube-api-access-mzmrt\") pod \"nova-api-0\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " pod="openstack/nova-api-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.901452 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.940235 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:36 crc kubenswrapper[4677]: I1203 14:12:36.959969 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.068683 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kjshq"] Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.130552 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kjshq" event={"ID":"fed8b246-9742-4b78-a630-b820a90ec8c8","Type":"ContainerStarted","Data":"edd0ba9d9a8bf831e07c95750a4530bae74135c3fa451797d0e00ca51e097ae9"} Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.234344 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:12:37 crc kubenswrapper[4677]: W1203 14:12:37.418902 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d2e8e2_6c98_4c30_9c6f_0817c401ed0b.slice/crio-e81b0517ea90e3772c9d3a168a626ef7e54c8edb1c5df7f4d4e1a4da5868e3bf WatchSource:0}: Error finding container e81b0517ea90e3772c9d3a168a626ef7e54c8edb1c5df7f4d4e1a4da5868e3bf: Status 404 returned error can't find the container with id e81b0517ea90e3772c9d3a168a626ef7e54c8edb1c5df7f4d4e1a4da5868e3bf Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.590691 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9vwd9"] Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.591917 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.595729 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.595903 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.622070 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9vwd9"] Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.666827 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.729386 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvhs6\" (UniqueName: \"kubernetes.io/projected/42fd76ce-441e-4c60-9220-89d31123f933-kube-api-access-hvhs6\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.729454 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-scripts\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.729632 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.729670 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-config-data\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.833985 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvhs6\" (UniqueName: \"kubernetes.io/projected/42fd76ce-441e-4c60-9220-89d31123f933-kube-api-access-hvhs6\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.834353 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-scripts\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.834543 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.834583 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-config-data\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.841405 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-scripts\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.845695 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-config-data\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.853006 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvhs6\" (UniqueName: \"kubernetes.io/projected/42fd76ce-441e-4c60-9220-89d31123f933-kube-api-access-hvhs6\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.873970 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9vwd9\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.883267 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.913517 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86755f6c8c-s8sbl"] Dec 03 14:12:37 crc kubenswrapper[4677]: I1203 14:12:37.953351 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.038127 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:38 crc kubenswrapper[4677]: W1203 14:12:38.069121 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58617ccc_ce55_4b21_816e_6a436b52b1a2.slice/crio-2391367bed3c7c558b1c51e0aafaf3f366e4fba758d72b9dc84b444375eaa3d6 WatchSource:0}: Error finding container 2391367bed3c7c558b1c51e0aafaf3f366e4fba758d72b9dc84b444375eaa3d6: Status 404 returned error can't find the container with id 2391367bed3c7c558b1c51e0aafaf3f366e4fba758d72b9dc84b444375eaa3d6 Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.168650 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kjshq" event={"ID":"fed8b246-9742-4b78-a630-b820a90ec8c8","Type":"ContainerStarted","Data":"580f81c7408191a5eb21928675151e5550399ebe1ed8cbbe07a3f0aa9034385c"} Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.172877 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37e0ed04-8f63-4fb6-9598-a4199b4b7258","Type":"ContainerStarted","Data":"555a4b248705cb3236bf1ca704579c60b675f018d31985e411533cb62e98d3c6"} Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.178149 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b","Type":"ContainerStarted","Data":"e81b0517ea90e3772c9d3a168a626ef7e54c8edb1c5df7f4d4e1a4da5868e3bf"} Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.182087 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" event={"ID":"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf","Type":"ContainerStarted","Data":"bacc73612c7259bf1b472d4107007591a25c8bf1f9fe6119afe04034402219be"} Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.183223 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c94b290-b422-470d-bf5d-f8dd6d28364d","Type":"ContainerStarted","Data":"b59d28a9d277df1102f52d69fe9bb48dc3fbcac58afefe5250340bb9d80847ac"} Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.196388 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-kjshq" podStartSLOduration=3.1963650120000002 podStartE2EDuration="3.196365012s" podCreationTimestamp="2025-12-03 14:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:38.190353084 +0000 UTC m=+1548.936685539" watchObservedRunningTime="2025-12-03 14:12:38.196365012 +0000 UTC m=+1548.942697467" Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.216825 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58617ccc-ce55-4b21-816e-6a436b52b1a2","Type":"ContainerStarted","Data":"2391367bed3c7c558b1c51e0aafaf3f366e4fba758d72b9dc84b444375eaa3d6"} Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.437605 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.437675 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:12:38 crc kubenswrapper[4677]: I1203 14:12:38.491739 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9vwd9"] Dec 03 14:12:39 crc kubenswrapper[4677]: I1203 14:12:39.234556 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerID="e2eb609ad114f74e2cf57387fbc15bf916dac81b877cd1a78780261721ebca68" exitCode=0 Dec 03 14:12:39 crc kubenswrapper[4677]: I1203 14:12:39.234677 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" event={"ID":"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf","Type":"ContainerDied","Data":"e2eb609ad114f74e2cf57387fbc15bf916dac81b877cd1a78780261721ebca68"} Dec 03 14:12:40 crc kubenswrapper[4677]: I1203 14:12:40.013148 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:40 crc kubenswrapper[4677]: I1203 14:12:40.050813 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:12:40 crc kubenswrapper[4677]: I1203 14:12:40.245601 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" event={"ID":"42fd76ce-441e-4c60-9220-89d31123f933","Type":"ContainerStarted","Data":"5ce98c44ab3b42e5da7ac3c35999e2877c64c754c66d128a05ac61c19239a863"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.313444 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b","Type":"ContainerStarted","Data":"bd4cf25b1ae6dbc951dabe951610ae35ea15903d469d83a681b551add1ff2a2f"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.313892 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bd4cf25b1ae6dbc951dabe951610ae35ea15903d469d83a681b551add1ff2a2f" gracePeriod=30 Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.338930 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" event={"ID":"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf","Type":"ContainerStarted","Data":"3d58c3e477a90f3bd83a928441860da1400291befd46e8aba7c4d491347e6e6f"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.339732 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.352384 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c94b290-b422-470d-bf5d-f8dd6d28364d","Type":"ContainerStarted","Data":"a1a1b69dfb2bba22c132554d8465f8181d225f032961bf190cd36b8d1ed3b8b2"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.361860 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.218622135 podStartE2EDuration="5.361833505s" podCreationTimestamp="2025-12-03 14:12:36 +0000 UTC" firstStartedPulling="2025-12-03 14:12:37.437099512 +0000 UTC m=+1548.183431967" lastFinishedPulling="2025-12-03 14:12:40.580310882 +0000 UTC m=+1551.326643337" observedRunningTime="2025-12-03 14:12:41.340192009 +0000 UTC m=+1552.086524464" watchObservedRunningTime="2025-12-03 14:12:41.361833505 +0000 UTC m=+1552.108165960" Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.375981 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58617ccc-ce55-4b21-816e-6a436b52b1a2","Type":"ContainerStarted","Data":"9f218e3bf39003c2500d0009ff62fa457938735cead3ba5bc7db7fe7df0a38e9"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.376028 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58617ccc-ce55-4b21-816e-6a436b52b1a2","Type":"ContainerStarted","Data":"bcc87ba7e49eaa43cd87d41ad4a5793031ce1a5cea45c5710256b929cd20c099"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.379649 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" podStartSLOduration=5.379629821 podStartE2EDuration="5.379629821s" podCreationTimestamp="2025-12-03 14:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:41.363600391 +0000 UTC m=+1552.109932856" watchObservedRunningTime="2025-12-03 14:12:41.379629821 +0000 UTC m=+1552.125962276" Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.401017 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" event={"ID":"42fd76ce-441e-4c60-9220-89d31123f933","Type":"ContainerStarted","Data":"2f225efbd9d3e9109b5f8d382b1e486ab86ba270b109f95891c04c3490e486f4"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.402975 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.895424206 podStartE2EDuration="5.40293889s" podCreationTimestamp="2025-12-03 14:12:36 +0000 UTC" firstStartedPulling="2025-12-03 14:12:38.081781793 +0000 UTC m=+1548.828114248" lastFinishedPulling="2025-12-03 14:12:40.589296477 +0000 UTC m=+1551.335628932" observedRunningTime="2025-12-03 14:12:41.398467274 +0000 UTC m=+1552.144799739" watchObservedRunningTime="2025-12-03 14:12:41.40293889 +0000 UTC m=+1552.149271345" Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.406684 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37e0ed04-8f63-4fb6-9598-a4199b4b7258","Type":"ContainerStarted","Data":"dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c"} Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.425810 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" podStartSLOduration=4.425785818 podStartE2EDuration="4.425785818s" podCreationTimestamp="2025-12-03 14:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:41.418246622 +0000 UTC m=+1552.164579077" watchObservedRunningTime="2025-12-03 14:12:41.425785818 +0000 UTC m=+1552.172118273" Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.438811 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.542934891 podStartE2EDuration="5.438788759s" podCreationTimestamp="2025-12-03 14:12:36 +0000 UTC" firstStartedPulling="2025-12-03 14:12:37.685681916 +0000 UTC m=+1548.432014371" lastFinishedPulling="2025-12-03 14:12:40.581535784 +0000 UTC m=+1551.327868239" observedRunningTime="2025-12-03 14:12:41.434941779 +0000 UTC m=+1552.181274234" watchObservedRunningTime="2025-12-03 14:12:41.438788759 +0000 UTC m=+1552.185121214" Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.580497 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:12:41 crc kubenswrapper[4677]: I1203 14:12:41.640340 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 14:12:42 crc kubenswrapper[4677]: I1203 14:12:42.417815 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c94b290-b422-470d-bf5d-f8dd6d28364d","Type":"ContainerStarted","Data":"4a61e61e88202d5376b223531762977cf7f45a57bfcca8dc7971bc4250fa2687"} Dec 03 14:12:42 crc kubenswrapper[4677]: I1203 14:12:42.418899 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-log" containerID="cri-o://a1a1b69dfb2bba22c132554d8465f8181d225f032961bf190cd36b8d1ed3b8b2" gracePeriod=30 Dec 03 14:12:42 crc kubenswrapper[4677]: I1203 14:12:42.419482 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-metadata" containerID="cri-o://4a61e61e88202d5376b223531762977cf7f45a57bfcca8dc7971bc4250fa2687" gracePeriod=30 Dec 03 14:12:42 crc kubenswrapper[4677]: I1203 14:12:42.446395 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.825864867 podStartE2EDuration="6.446332417s" podCreationTimestamp="2025-12-03 14:12:36 +0000 UTC" firstStartedPulling="2025-12-03 14:12:37.914104645 +0000 UTC m=+1548.660437100" lastFinishedPulling="2025-12-03 14:12:40.534572195 +0000 UTC m=+1551.280904650" observedRunningTime="2025-12-03 14:12:42.44225915 +0000 UTC m=+1553.188591625" watchObservedRunningTime="2025-12-03 14:12:42.446332417 +0000 UTC m=+1553.192664892" Dec 03 14:12:43 crc kubenswrapper[4677]: E1203 14:12:43.369490 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice/crio-1d01743f5632ec817b7e11e68f4ff78583597cd7be455d14fa7f0095553f2991\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice/crio-5820cbc2e989810a5d19c0408aa224a19ea654893fb5c2a361f17b4c04bf9f22\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3227517_ddda_40a6_b219_420ca2b9a471.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d232644_1f01_4ed4_9ca5_d4259a0d1c99.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.430897 4677 generic.go:334] "Generic (PLEG): container finished" podID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerID="4a61e61e88202d5376b223531762977cf7f45a57bfcca8dc7971bc4250fa2687" exitCode=0 Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.430940 4677 generic.go:334] "Generic (PLEG): container finished" podID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerID="a1a1b69dfb2bba22c132554d8465f8181d225f032961bf190cd36b8d1ed3b8b2" exitCode=143 Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.431891 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c94b290-b422-470d-bf5d-f8dd6d28364d","Type":"ContainerDied","Data":"4a61e61e88202d5376b223531762977cf7f45a57bfcca8dc7971bc4250fa2687"} Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.431922 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c94b290-b422-470d-bf5d-f8dd6d28364d","Type":"ContainerDied","Data":"a1a1b69dfb2bba22c132554d8465f8181d225f032961bf190cd36b8d1ed3b8b2"} Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.431935 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c94b290-b422-470d-bf5d-f8dd6d28364d","Type":"ContainerDied","Data":"b59d28a9d277df1102f52d69fe9bb48dc3fbcac58afefe5250340bb9d80847ac"} Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.431968 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b59d28a9d277df1102f52d69fe9bb48dc3fbcac58afefe5250340bb9d80847ac" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.519829 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.643643 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rxg8\" (UniqueName: \"kubernetes.io/projected/7c94b290-b422-470d-bf5d-f8dd6d28364d-kube-api-access-7rxg8\") pod \"7c94b290-b422-470d-bf5d-f8dd6d28364d\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.644246 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-config-data\") pod \"7c94b290-b422-470d-bf5d-f8dd6d28364d\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.644508 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c94b290-b422-470d-bf5d-f8dd6d28364d-logs\") pod \"7c94b290-b422-470d-bf5d-f8dd6d28364d\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.644573 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-combined-ca-bundle\") pod \"7c94b290-b422-470d-bf5d-f8dd6d28364d\" (UID: \"7c94b290-b422-470d-bf5d-f8dd6d28364d\") " Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.645506 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c94b290-b422-470d-bf5d-f8dd6d28364d-logs" (OuterVolumeSpecName: "logs") pod "7c94b290-b422-470d-bf5d-f8dd6d28364d" (UID: "7c94b290-b422-470d-bf5d-f8dd6d28364d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.645801 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c94b290-b422-470d-bf5d-f8dd6d28364d-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.669150 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c94b290-b422-470d-bf5d-f8dd6d28364d-kube-api-access-7rxg8" (OuterVolumeSpecName: "kube-api-access-7rxg8") pod "7c94b290-b422-470d-bf5d-f8dd6d28364d" (UID: "7c94b290-b422-470d-bf5d-f8dd6d28364d"). InnerVolumeSpecName "kube-api-access-7rxg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.677717 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-config-data" (OuterVolumeSpecName: "config-data") pod "7c94b290-b422-470d-bf5d-f8dd6d28364d" (UID: "7c94b290-b422-470d-bf5d-f8dd6d28364d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.681715 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c94b290-b422-470d-bf5d-f8dd6d28364d" (UID: "7c94b290-b422-470d-bf5d-f8dd6d28364d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.747500 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rxg8\" (UniqueName: \"kubernetes.io/projected/7c94b290-b422-470d-bf5d-f8dd6d28364d-kube-api-access-7rxg8\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.747536 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:43 crc kubenswrapper[4677]: I1203 14:12:43.747548 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c94b290-b422-470d-bf5d-f8dd6d28364d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.440420 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.471742 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.487790 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.503594 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:44 crc kubenswrapper[4677]: E1203 14:12:44.504457 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-metadata" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.504484 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-metadata" Dec 03 14:12:44 crc kubenswrapper[4677]: E1203 14:12:44.504506 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-log" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.504514 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-log" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.504785 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-metadata" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.504833 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" containerName="nova-metadata-log" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.506443 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.520555 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.521166 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.529679 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.567407 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqnhn\" (UniqueName: \"kubernetes.io/projected/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-kube-api-access-kqnhn\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.567575 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-logs\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.567720 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.567786 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.568082 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-config-data\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.670394 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.670843 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.671151 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-config-data\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.671244 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqnhn\" (UniqueName: \"kubernetes.io/projected/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-kube-api-access-kqnhn\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.671313 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-logs\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.672125 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-logs\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.678886 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.681601 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.692736 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-config-data\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.693739 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqnhn\" (UniqueName: \"kubernetes.io/projected/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-kube-api-access-kqnhn\") pod \"nova-metadata-0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " pod="openstack/nova-metadata-0" Dec 03 14:12:44 crc kubenswrapper[4677]: I1203 14:12:44.825525 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:45 crc kubenswrapper[4677]: I1203 14:12:45.369553 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:45 crc kubenswrapper[4677]: I1203 14:12:45.453019 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0","Type":"ContainerStarted","Data":"2c0993e5a279ca95482ad02bb39058205796bdd0bb3f5ba689e2132ed10e1354"} Dec 03 14:12:45 crc kubenswrapper[4677]: I1203 14:12:45.990297 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c94b290-b422-470d-bf5d-f8dd6d28364d" path="/var/lib/kubelet/pods/7c94b290-b422-470d-bf5d-f8dd6d28364d/volumes" Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.467057 4677 generic.go:334] "Generic (PLEG): container finished" podID="fed8b246-9742-4b78-a630-b820a90ec8c8" containerID="580f81c7408191a5eb21928675151e5550399ebe1ed8cbbe07a3f0aa9034385c" exitCode=0 Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.467132 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kjshq" event={"ID":"fed8b246-9742-4b78-a630-b820a90ec8c8","Type":"ContainerDied","Data":"580f81c7408191a5eb21928675151e5550399ebe1ed8cbbe07a3f0aa9034385c"} Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.470461 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0","Type":"ContainerStarted","Data":"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59"} Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.470523 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0","Type":"ContainerStarted","Data":"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6"} Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.525522 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.525500193 podStartE2EDuration="2.525500193s" podCreationTimestamp="2025-12-03 14:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:46.511485076 +0000 UTC m=+1557.257817531" watchObservedRunningTime="2025-12-03 14:12:46.525500193 +0000 UTC m=+1557.271832658" Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.640273 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.683386 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.942301 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.960318 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:12:46 crc kubenswrapper[4677]: I1203 14:12:46.960373 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.036207 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f955c7f4c-hfdt9"] Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.036589 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" podUID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerName="dnsmasq-dns" containerID="cri-o://b64ed340e3bc0b3fa5528dc8e8f644e44880f1b110f59e8974d6330b55aeff0a" gracePeriod=10 Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.482345 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerStarted","Data":"8730132cc6488e77198dfa36fa1618e12ece424d3757ff62a0668cbcd1f63037"} Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.483055 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.484941 4677 generic.go:334] "Generic (PLEG): container finished" podID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerID="b64ed340e3bc0b3fa5528dc8e8f644e44880f1b110f59e8974d6330b55aeff0a" exitCode=0 Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.485156 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" event={"ID":"8e1c647c-16bc-4421-9bc8-7395212da3fe","Type":"ContainerDied","Data":"b64ed340e3bc0b3fa5528dc8e8f644e44880f1b110f59e8974d6330b55aeff0a"} Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.513793 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.827746411 podStartE2EDuration="28.513769157s" podCreationTimestamp="2025-12-03 14:12:19 +0000 UTC" firstStartedPulling="2025-12-03 14:12:20.831866747 +0000 UTC m=+1531.578199202" lastFinishedPulling="2025-12-03 14:12:46.517889483 +0000 UTC m=+1557.264221948" observedRunningTime="2025-12-03 14:12:47.509850694 +0000 UTC m=+1558.256183179" watchObservedRunningTime="2025-12-03 14:12:47.513769157 +0000 UTC m=+1558.260101632" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.534404 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.628673 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.764219 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-swift-storage-0\") pod \"8e1c647c-16bc-4421-9bc8-7395212da3fe\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.764337 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt8df\" (UniqueName: \"kubernetes.io/projected/8e1c647c-16bc-4421-9bc8-7395212da3fe-kube-api-access-jt8df\") pod \"8e1c647c-16bc-4421-9bc8-7395212da3fe\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.764392 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-svc\") pod \"8e1c647c-16bc-4421-9bc8-7395212da3fe\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.764552 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-config\") pod \"8e1c647c-16bc-4421-9bc8-7395212da3fe\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.764615 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-nb\") pod \"8e1c647c-16bc-4421-9bc8-7395212da3fe\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.764646 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-sb\") pod \"8e1c647c-16bc-4421-9bc8-7395212da3fe\" (UID: \"8e1c647c-16bc-4421-9bc8-7395212da3fe\") " Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.772737 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e1c647c-16bc-4421-9bc8-7395212da3fe-kube-api-access-jt8df" (OuterVolumeSpecName: "kube-api-access-jt8df") pod "8e1c647c-16bc-4421-9bc8-7395212da3fe" (UID: "8e1c647c-16bc-4421-9bc8-7395212da3fe"). InnerVolumeSpecName "kube-api-access-jt8df". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.828936 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-config" (OuterVolumeSpecName: "config") pod "8e1c647c-16bc-4421-9bc8-7395212da3fe" (UID: "8e1c647c-16bc-4421-9bc8-7395212da3fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.836119 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e1c647c-16bc-4421-9bc8-7395212da3fe" (UID: "8e1c647c-16bc-4421-9bc8-7395212da3fe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.867370 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt8df\" (UniqueName: \"kubernetes.io/projected/8e1c647c-16bc-4421-9bc8-7395212da3fe-kube-api-access-jt8df\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.867397 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.867406 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.869488 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e1c647c-16bc-4421-9bc8-7395212da3fe" (UID: "8e1c647c-16bc-4421-9bc8-7395212da3fe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.872655 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e1c647c-16bc-4421-9bc8-7395212da3fe" (UID: "8e1c647c-16bc-4421-9bc8-7395212da3fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.874179 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8e1c647c-16bc-4421-9bc8-7395212da3fe" (UID: "8e1c647c-16bc-4421-9bc8-7395212da3fe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.906464 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.974457 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.974493 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:47 crc kubenswrapper[4677]: I1203 14:12:47.974504 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1c647c-16bc-4421-9bc8-7395212da3fe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.042101 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.042101 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.075467 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nljkw\" (UniqueName: \"kubernetes.io/projected/fed8b246-9742-4b78-a630-b820a90ec8c8-kube-api-access-nljkw\") pod \"fed8b246-9742-4b78-a630-b820a90ec8c8\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.075526 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-combined-ca-bundle\") pod \"fed8b246-9742-4b78-a630-b820a90ec8c8\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.075761 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-scripts\") pod \"fed8b246-9742-4b78-a630-b820a90ec8c8\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.075876 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-config-data\") pod \"fed8b246-9742-4b78-a630-b820a90ec8c8\" (UID: \"fed8b246-9742-4b78-a630-b820a90ec8c8\") " Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.081292 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-scripts" (OuterVolumeSpecName: "scripts") pod "fed8b246-9742-4b78-a630-b820a90ec8c8" (UID: "fed8b246-9742-4b78-a630-b820a90ec8c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.081335 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fed8b246-9742-4b78-a630-b820a90ec8c8-kube-api-access-nljkw" (OuterVolumeSpecName: "kube-api-access-nljkw") pod "fed8b246-9742-4b78-a630-b820a90ec8c8" (UID: "fed8b246-9742-4b78-a630-b820a90ec8c8"). InnerVolumeSpecName "kube-api-access-nljkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.107056 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-config-data" (OuterVolumeSpecName: "config-data") pod "fed8b246-9742-4b78-a630-b820a90ec8c8" (UID: "fed8b246-9742-4b78-a630-b820a90ec8c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.107977 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fed8b246-9742-4b78-a630-b820a90ec8c8" (UID: "fed8b246-9742-4b78-a630-b820a90ec8c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.180249 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.180285 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nljkw\" (UniqueName: \"kubernetes.io/projected/fed8b246-9742-4b78-a630-b820a90ec8c8-kube-api-access-nljkw\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.180297 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.180306 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fed8b246-9742-4b78-a630-b820a90ec8c8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.495185 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kjshq" event={"ID":"fed8b246-9742-4b78-a630-b820a90ec8c8","Type":"ContainerDied","Data":"edd0ba9d9a8bf831e07c95750a4530bae74135c3fa451797d0e00ca51e097ae9"} Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.495225 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edd0ba9d9a8bf831e07c95750a4530bae74135c3fa451797d0e00ca51e097ae9" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.495286 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kjshq" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.503406 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" event={"ID":"8e1c647c-16bc-4421-9bc8-7395212da3fe","Type":"ContainerDied","Data":"60157941364cb6f27e072665c5c08ddb24c7252112c2775762c7cec9938e3138"} Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.503495 4677 scope.go:117] "RemoveContainer" containerID="b64ed340e3bc0b3fa5528dc8e8f644e44880f1b110f59e8974d6330b55aeff0a" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.503743 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f955c7f4c-hfdt9" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.542527 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f955c7f4c-hfdt9"] Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.543865 4677 scope.go:117] "RemoveContainer" containerID="f730497aa99e9a0c54fcfbb2cb7a0fb6913270a37dbdb39e9acb7ad5edadc3df" Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.572460 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f955c7f4c-hfdt9"] Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.676019 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.676292 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-log" containerID="cri-o://bcc87ba7e49eaa43cd87d41ad4a5793031ce1a5cea45c5710256b929cd20c099" gracePeriod=30 Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.676382 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-api" containerID="cri-o://9f218e3bf39003c2500d0009ff62fa457938735cead3ba5bc7db7fe7df0a38e9" gracePeriod=30 Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.694277 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.710005 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.710257 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-log" containerID="cri-o://883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6" gracePeriod=30 Dec 03 14:12:48 crc kubenswrapper[4677]: I1203 14:12:48.710417 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-metadata" containerID="cri-o://50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59" gracePeriod=30 Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.313250 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.501149 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-config-data\") pod \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.501397 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqnhn\" (UniqueName: \"kubernetes.io/projected/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-kube-api-access-kqnhn\") pod \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.501429 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-combined-ca-bundle\") pod \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.501459 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-logs\") pod \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.501487 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-nova-metadata-tls-certs\") pod \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\" (UID: \"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0\") " Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.501824 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-logs" (OuterVolumeSpecName: "logs") pod "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" (UID: "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.513302 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-kube-api-access-kqnhn" (OuterVolumeSpecName: "kube-api-access-kqnhn") pod "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" (UID: "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0"). InnerVolumeSpecName "kube-api-access-kqnhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.548165 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-config-data" (OuterVolumeSpecName: "config-data") pod "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" (UID: "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.568133 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" (UID: "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.570119 4677 generic.go:334] "Generic (PLEG): container finished" podID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerID="50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59" exitCode=0 Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.570168 4677 generic.go:334] "Generic (PLEG): container finished" podID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerID="883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6" exitCode=143 Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.570219 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0","Type":"ContainerDied","Data":"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59"} Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.570247 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0","Type":"ContainerDied","Data":"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6"} Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.570257 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca0eabc1-86f9-48ce-9cb5-aa9c964702b0","Type":"ContainerDied","Data":"2c0993e5a279ca95482ad02bb39058205796bdd0bb3f5ba689e2132ed10e1354"} Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.570272 4677 scope.go:117] "RemoveContainer" containerID="50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.570205 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.575231 4677 generic.go:334] "Generic (PLEG): container finished" podID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerID="bcc87ba7e49eaa43cd87d41ad4a5793031ce1a5cea45c5710256b929cd20c099" exitCode=143 Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.575311 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58617ccc-ce55-4b21-816e-6a436b52b1a2","Type":"ContainerDied","Data":"bcc87ba7e49eaa43cd87d41ad4a5793031ce1a5cea45c5710256b929cd20c099"} Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.575401 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="37e0ed04-8f63-4fb6-9598-a4199b4b7258" containerName="nova-scheduler-scheduler" containerID="cri-o://dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" gracePeriod=30 Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.582116 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" (UID: "ca0eabc1-86f9-48ce-9cb5-aa9c964702b0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.604110 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqnhn\" (UniqueName: \"kubernetes.io/projected/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-kube-api-access-kqnhn\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.604151 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.604162 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.604174 4677 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.604185 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.678123 4677 scope.go:117] "RemoveContainer" containerID="883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.702286 4677 scope.go:117] "RemoveContainer" containerID="50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59" Dec 03 14:12:49 crc kubenswrapper[4677]: E1203 14:12:49.702708 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59\": container with ID starting with 50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59 not found: ID does not exist" containerID="50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.702740 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59"} err="failed to get container status \"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59\": rpc error: code = NotFound desc = could not find container \"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59\": container with ID starting with 50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59 not found: ID does not exist" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.702761 4677 scope.go:117] "RemoveContainer" containerID="883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6" Dec 03 14:12:49 crc kubenswrapper[4677]: E1203 14:12:49.703373 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6\": container with ID starting with 883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6 not found: ID does not exist" containerID="883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.703443 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6"} err="failed to get container status \"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6\": rpc error: code = NotFound desc = could not find container \"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6\": container with ID starting with 883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6 not found: ID does not exist" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.703494 4677 scope.go:117] "RemoveContainer" containerID="50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.703884 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59"} err="failed to get container status \"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59\": rpc error: code = NotFound desc = could not find container \"50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59\": container with ID starting with 50e9966fa9bbf6d48356bb9f441e32b117cc0b8391cacba320f4179ca2326e59 not found: ID does not exist" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.703927 4677 scope.go:117] "RemoveContainer" containerID="883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.704253 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6"} err="failed to get container status \"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6\": rpc error: code = NotFound desc = could not find container \"883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6\": container with ID starting with 883b312a152e919c0d915ed570fcb28a5083180099c46eeead1941dfe6cbfac6 not found: ID does not exist" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.903247 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.911347 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.946778 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:49 crc kubenswrapper[4677]: E1203 14:12:49.947808 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerName="init" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.947841 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerName="init" Dec 03 14:12:49 crc kubenswrapper[4677]: E1203 14:12:49.947856 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed8b246-9742-4b78-a630-b820a90ec8c8" containerName="nova-manage" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.947876 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed8b246-9742-4b78-a630-b820a90ec8c8" containerName="nova-manage" Dec 03 14:12:49 crc kubenswrapper[4677]: E1203 14:12:49.947904 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-log" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.947918 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-log" Dec 03 14:12:49 crc kubenswrapper[4677]: E1203 14:12:49.947986 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-metadata" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.947999 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-metadata" Dec 03 14:12:49 crc kubenswrapper[4677]: E1203 14:12:49.948072 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerName="dnsmasq-dns" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.948083 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerName="dnsmasq-dns" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.948806 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e1c647c-16bc-4421-9bc8-7395212da3fe" containerName="dnsmasq-dns" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.948874 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed8b246-9742-4b78-a630-b820a90ec8c8" containerName="nova-manage" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.948909 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-log" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.948938 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" containerName="nova-metadata-metadata" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.971296 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.971419 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.977393 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 14:12:49 crc kubenswrapper[4677]: I1203 14:12:49.978878 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.013197 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e1c647c-16bc-4421-9bc8-7395212da3fe" path="/var/lib/kubelet/pods/8e1c647c-16bc-4421-9bc8-7395212da3fe/volumes" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.014978 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca0eabc1-86f9-48ce-9cb5-aa9c964702b0" path="/var/lib/kubelet/pods/ca0eabc1-86f9-48ce-9cb5-aa9c964702b0/volumes" Dec 03 14:12:50 crc kubenswrapper[4677]: E1203 14:12:50.038284 4677 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/c8f26c1a5aa54f64c9443780e2a3a6325720691c7764744e7e7eed886fcb404d/diff" to get inode usage: stat /var/lib/containers/storage/overlay/c8f26c1a5aa54f64c9443780e2a3a6325720691c7764744e7e7eed886fcb404d/diff: no such file or directory, extraDiskErr: Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.112380 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57mhb\" (UniqueName: \"kubernetes.io/projected/7e74337e-3593-494a-8a23-1b787d670087-kube-api-access-57mhb\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.112431 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-config-data\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.112726 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e74337e-3593-494a-8a23-1b787d670087-logs\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.112775 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.113022 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.215409 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.215554 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57mhb\" (UniqueName: \"kubernetes.io/projected/7e74337e-3593-494a-8a23-1b787d670087-kube-api-access-57mhb\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.215592 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-config-data\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.215655 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e74337e-3593-494a-8a23-1b787d670087-logs\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.215701 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.216230 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e74337e-3593-494a-8a23-1b787d670087-logs\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.220906 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.220917 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-config-data\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.224142 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.234437 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57mhb\" (UniqueName: \"kubernetes.io/projected/7e74337e-3593-494a-8a23-1b787d670087-kube-api-access-57mhb\") pod \"nova-metadata-0\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.299708 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:12:50 crc kubenswrapper[4677]: I1203 14:12:50.799392 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:12:51 crc kubenswrapper[4677]: I1203 14:12:51.609165 4677 generic.go:334] "Generic (PLEG): container finished" podID="42fd76ce-441e-4c60-9220-89d31123f933" containerID="2f225efbd9d3e9109b5f8d382b1e486ab86ba270b109f95891c04c3490e486f4" exitCode=0 Dec 03 14:12:51 crc kubenswrapper[4677]: I1203 14:12:51.609271 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" event={"ID":"42fd76ce-441e-4c60-9220-89d31123f933","Type":"ContainerDied","Data":"2f225efbd9d3e9109b5f8d382b1e486ab86ba270b109f95891c04c3490e486f4"} Dec 03 14:12:51 crc kubenswrapper[4677]: I1203 14:12:51.612024 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e74337e-3593-494a-8a23-1b787d670087","Type":"ContainerStarted","Data":"aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6"} Dec 03 14:12:51 crc kubenswrapper[4677]: I1203 14:12:51.612055 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e74337e-3593-494a-8a23-1b787d670087","Type":"ContainerStarted","Data":"95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e"} Dec 03 14:12:51 crc kubenswrapper[4677]: I1203 14:12:51.612068 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e74337e-3593-494a-8a23-1b787d670087","Type":"ContainerStarted","Data":"194347dc3e32c35785a07e09b47927c396010442097688ae3c360717f86ab7dc"} Dec 03 14:12:51 crc kubenswrapper[4677]: E1203 14:12:51.642684 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:12:51 crc kubenswrapper[4677]: E1203 14:12:51.645202 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:12:51 crc kubenswrapper[4677]: E1203 14:12:51.648567 4677 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:12:51 crc kubenswrapper[4677]: E1203 14:12:51.648639 4677 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="37e0ed04-8f63-4fb6-9598-a4199b4b7258" containerName="nova-scheduler-scheduler" Dec 03 14:12:51 crc kubenswrapper[4677]: I1203 14:12:51.657123 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.657104691 podStartE2EDuration="2.657104691s" podCreationTimestamp="2025-12-03 14:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:51.650791081 +0000 UTC m=+1562.397123546" watchObservedRunningTime="2025-12-03 14:12:51.657104691 +0000 UTC m=+1562.403437166" Dec 03 14:12:52 crc kubenswrapper[4677]: I1203 14:12:52.631571 4677 generic.go:334] "Generic (PLEG): container finished" podID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerID="9f218e3bf39003c2500d0009ff62fa457938735cead3ba5bc7db7fe7df0a38e9" exitCode=0 Dec 03 14:12:52 crc kubenswrapper[4677]: I1203 14:12:52.631687 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58617ccc-ce55-4b21-816e-6a436b52b1a2","Type":"ContainerDied","Data":"9f218e3bf39003c2500d0009ff62fa457938735cead3ba5bc7db7fe7df0a38e9"} Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.152270 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.160883 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.199346 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294495 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58617ccc-ce55-4b21-816e-6a436b52b1a2-logs\") pod \"58617ccc-ce55-4b21-816e-6a436b52b1a2\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294570 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-combined-ca-bundle\") pod \"42fd76ce-441e-4c60-9220-89d31123f933\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294621 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-config-data\") pod \"42fd76ce-441e-4c60-9220-89d31123f933\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294691 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-combined-ca-bundle\") pod \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294730 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-combined-ca-bundle\") pod \"58617ccc-ce55-4b21-816e-6a436b52b1a2\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294776 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-config-data\") pod \"58617ccc-ce55-4b21-816e-6a436b52b1a2\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294810 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skrf2\" (UniqueName: \"kubernetes.io/projected/37e0ed04-8f63-4fb6-9598-a4199b4b7258-kube-api-access-skrf2\") pod \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294867 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzmrt\" (UniqueName: \"kubernetes.io/projected/58617ccc-ce55-4b21-816e-6a436b52b1a2-kube-api-access-mzmrt\") pod \"58617ccc-ce55-4b21-816e-6a436b52b1a2\" (UID: \"58617ccc-ce55-4b21-816e-6a436b52b1a2\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294891 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-config-data\") pod \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\" (UID: \"37e0ed04-8f63-4fb6-9598-a4199b4b7258\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294930 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvhs6\" (UniqueName: \"kubernetes.io/projected/42fd76ce-441e-4c60-9220-89d31123f933-kube-api-access-hvhs6\") pod \"42fd76ce-441e-4c60-9220-89d31123f933\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.294986 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-scripts\") pod \"42fd76ce-441e-4c60-9220-89d31123f933\" (UID: \"42fd76ce-441e-4c60-9220-89d31123f933\") " Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.295841 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58617ccc-ce55-4b21-816e-6a436b52b1a2-logs" (OuterVolumeSpecName: "logs") pod "58617ccc-ce55-4b21-816e-6a436b52b1a2" (UID: "58617ccc-ce55-4b21-816e-6a436b52b1a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.302704 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37e0ed04-8f63-4fb6-9598-a4199b4b7258-kube-api-access-skrf2" (OuterVolumeSpecName: "kube-api-access-skrf2") pod "37e0ed04-8f63-4fb6-9598-a4199b4b7258" (UID: "37e0ed04-8f63-4fb6-9598-a4199b4b7258"). InnerVolumeSpecName "kube-api-access-skrf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.303197 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-scripts" (OuterVolumeSpecName: "scripts") pod "42fd76ce-441e-4c60-9220-89d31123f933" (UID: "42fd76ce-441e-4c60-9220-89d31123f933"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.303499 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42fd76ce-441e-4c60-9220-89d31123f933-kube-api-access-hvhs6" (OuterVolumeSpecName: "kube-api-access-hvhs6") pod "42fd76ce-441e-4c60-9220-89d31123f933" (UID: "42fd76ce-441e-4c60-9220-89d31123f933"). InnerVolumeSpecName "kube-api-access-hvhs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.307311 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58617ccc-ce55-4b21-816e-6a436b52b1a2-kube-api-access-mzmrt" (OuterVolumeSpecName: "kube-api-access-mzmrt") pod "58617ccc-ce55-4b21-816e-6a436b52b1a2" (UID: "58617ccc-ce55-4b21-816e-6a436b52b1a2"). InnerVolumeSpecName "kube-api-access-mzmrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.328988 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58617ccc-ce55-4b21-816e-6a436b52b1a2" (UID: "58617ccc-ce55-4b21-816e-6a436b52b1a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.330757 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-config-data" (OuterVolumeSpecName: "config-data") pod "37e0ed04-8f63-4fb6-9598-a4199b4b7258" (UID: "37e0ed04-8f63-4fb6-9598-a4199b4b7258"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.331801 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37e0ed04-8f63-4fb6-9598-a4199b4b7258" (UID: "37e0ed04-8f63-4fb6-9598-a4199b4b7258"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.337257 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-config-data" (OuterVolumeSpecName: "config-data") pod "42fd76ce-441e-4c60-9220-89d31123f933" (UID: "42fd76ce-441e-4c60-9220-89d31123f933"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.337559 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42fd76ce-441e-4c60-9220-89d31123f933" (UID: "42fd76ce-441e-4c60-9220-89d31123f933"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.338286 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-config-data" (OuterVolumeSpecName: "config-data") pod "58617ccc-ce55-4b21-816e-6a436b52b1a2" (UID: "58617ccc-ce55-4b21-816e-6a436b52b1a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397287 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397333 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58617ccc-ce55-4b21-816e-6a436b52b1a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397346 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skrf2\" (UniqueName: \"kubernetes.io/projected/37e0ed04-8f63-4fb6-9598-a4199b4b7258-kube-api-access-skrf2\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397362 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzmrt\" (UniqueName: \"kubernetes.io/projected/58617ccc-ce55-4b21-816e-6a436b52b1a2-kube-api-access-mzmrt\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397375 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397388 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvhs6\" (UniqueName: \"kubernetes.io/projected/42fd76ce-441e-4c60-9220-89d31123f933-kube-api-access-hvhs6\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397399 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397409 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58617ccc-ce55-4b21-816e-6a436b52b1a2-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397420 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397428 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fd76ce-441e-4c60-9220-89d31123f933-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.397436 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e0ed04-8f63-4fb6-9598-a4199b4b7258-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.647406 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58617ccc-ce55-4b21-816e-6a436b52b1a2","Type":"ContainerDied","Data":"2391367bed3c7c558b1c51e0aafaf3f366e4fba758d72b9dc84b444375eaa3d6"} Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.647462 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.648571 4677 scope.go:117] "RemoveContainer" containerID="9f218e3bf39003c2500d0009ff62fa457938735cead3ba5bc7db7fe7df0a38e9" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.651520 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" event={"ID":"42fd76ce-441e-4c60-9220-89d31123f933","Type":"ContainerDied","Data":"5ce98c44ab3b42e5da7ac3c35999e2877c64c754c66d128a05ac61c19239a863"} Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.651674 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ce98c44ab3b42e5da7ac3c35999e2877c64c754c66d128a05ac61c19239a863" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.651996 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9vwd9" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.655364 4677 generic.go:334] "Generic (PLEG): container finished" podID="37e0ed04-8f63-4fb6-9598-a4199b4b7258" containerID="dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" exitCode=0 Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.655421 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37e0ed04-8f63-4fb6-9598-a4199b4b7258","Type":"ContainerDied","Data":"dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c"} Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.655456 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37e0ed04-8f63-4fb6-9598-a4199b4b7258","Type":"ContainerDied","Data":"555a4b248705cb3236bf1ca704579c60b675f018d31985e411533cb62e98d3c6"} Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.655531 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.728826 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: E1203 14:12:53.729338 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42fd76ce-441e-4c60-9220-89d31123f933" containerName="nova-cell1-conductor-db-sync" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729365 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="42fd76ce-441e-4c60-9220-89d31123f933" containerName="nova-cell1-conductor-db-sync" Dec 03 14:12:53 crc kubenswrapper[4677]: E1203 14:12:53.729381 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-api" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729388 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-api" Dec 03 14:12:53 crc kubenswrapper[4677]: E1203 14:12:53.729416 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-log" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729423 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-log" Dec 03 14:12:53 crc kubenswrapper[4677]: E1203 14:12:53.729445 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e0ed04-8f63-4fb6-9598-a4199b4b7258" containerName="nova-scheduler-scheduler" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729453 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e0ed04-8f63-4fb6-9598-a4199b4b7258" containerName="nova-scheduler-scheduler" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729666 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="42fd76ce-441e-4c60-9220-89d31123f933" containerName="nova-cell1-conductor-db-sync" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729687 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-api" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729701 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" containerName="nova-api-log" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.729715 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e0ed04-8f63-4fb6-9598-a4199b4b7258" containerName="nova-scheduler-scheduler" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.730508 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.733935 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.770033 4677 scope.go:117] "RemoveContainer" containerID="bcc87ba7e49eaa43cd87d41ad4a5793031ce1a5cea45c5710256b929cd20c099" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.774081 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.814168 4677 scope.go:117] "RemoveContainer" containerID="dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.815975 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.816036 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.816081 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhwvk\" (UniqueName: \"kubernetes.io/projected/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-kube-api-access-dhwvk\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.816447 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.861759 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.875373 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.880010 4677 scope.go:117] "RemoveContainer" containerID="dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" Dec 03 14:12:53 crc kubenswrapper[4677]: E1203 14:12:53.880608 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c\": container with ID starting with dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c not found: ID does not exist" containerID="dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.880657 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c"} err="failed to get container status \"dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c\": rpc error: code = NotFound desc = could not find container \"dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c\": container with ID starting with dba18ec08d472789ba6423d47e9dd9f13f0a99759d029e7802191f643422c20c not found: ID does not exist" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.884103 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.892934 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.894919 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.897454 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.916390 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.917361 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.917419 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.917475 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.917533 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cktvk\" (UniqueName: \"kubernetes.io/projected/8d042f25-044e-431e-b353-9f37399f41f3-kube-api-access-cktvk\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.917561 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhwvk\" (UniqueName: \"kubernetes.io/projected/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-kube-api-access-dhwvk\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.917607 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-config-data\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.921833 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.926060 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.926774 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.928974 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.931187 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.935728 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhwvk\" (UniqueName: \"kubernetes.io/projected/1d4597ea-0a0a-4f5a-bae9-d69050c1fd56-kube-api-access-dhwvk\") pod \"nova-cell1-conductor-0\" (UID: \"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.936549 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.987079 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37e0ed04-8f63-4fb6-9598-a4199b4b7258" path="/var/lib/kubelet/pods/37e0ed04-8f63-4fb6-9598-a4199b4b7258/volumes" Dec 03 14:12:53 crc kubenswrapper[4677]: I1203 14:12:53.987614 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58617ccc-ce55-4b21-816e-6a436b52b1a2" path="/var/lib/kubelet/pods/58617ccc-ce55-4b21-816e-6a436b52b1a2/volumes" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.019180 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cktvk\" (UniqueName: \"kubernetes.io/projected/8d042f25-044e-431e-b353-9f37399f41f3-kube-api-access-cktvk\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.019262 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-config-data\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.019306 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0752718-309a-41d3-80af-563483da9f5b-logs\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.019322 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs7w9\" (UniqueName: \"kubernetes.io/projected/b0752718-309a-41d3-80af-563483da9f5b-kube-api-access-rs7w9\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.019341 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-config-data\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.019413 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.019481 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.026004 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.026122 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-config-data\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.038086 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cktvk\" (UniqueName: \"kubernetes.io/projected/8d042f25-044e-431e-b353-9f37399f41f3-kube-api-access-cktvk\") pod \"nova-scheduler-0\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " pod="openstack/nova-scheduler-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.084986 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.121195 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs7w9\" (UniqueName: \"kubernetes.io/projected/b0752718-309a-41d3-80af-563483da9f5b-kube-api-access-rs7w9\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.121247 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0752718-309a-41d3-80af-563483da9f5b-logs\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.121279 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-config-data\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.121467 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.122598 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0752718-309a-41d3-80af-563483da9f5b-logs\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.125202 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.125550 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-config-data\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.140304 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs7w9\" (UniqueName: \"kubernetes.io/projected/b0752718-309a-41d3-80af-563483da9f5b-kube-api-access-rs7w9\") pod \"nova-api-0\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.218208 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.428965 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.587388 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.671310 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56","Type":"ContainerStarted","Data":"9587efb4758e3c9dc1a89114af890a24a8d43b03b0a942321e29275d2d277951"} Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.707120 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:12:54 crc kubenswrapper[4677]: I1203 14:12:54.946903 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.306273 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.306338 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.686088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1d4597ea-0a0a-4f5a-bae9-d69050c1fd56","Type":"ContainerStarted","Data":"6a31eee20786cc98ed3b6ee9dcd77536289387de8e5e37e12eefca7557fc5b79"} Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.686820 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.689991 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d042f25-044e-431e-b353-9f37399f41f3","Type":"ContainerStarted","Data":"9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de"} Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.690025 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d042f25-044e-431e-b353-9f37399f41f3","Type":"ContainerStarted","Data":"8a804006adf53aa4a575a28f72d2c84d43b05a9b8f3a46133ca24dfb094cbbbf"} Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.696136 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b0752718-309a-41d3-80af-563483da9f5b","Type":"ContainerStarted","Data":"7a2fdf98a127eae0b332d0b76e17baa09257f3b5ab3cfe89e6ab4f6a9dd08810"} Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.696196 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b0752718-309a-41d3-80af-563483da9f5b","Type":"ContainerStarted","Data":"7e79b4c0cb29b438e3276028cc2551fa2016451bbc72028e590fa0f2d6985e09"} Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.696214 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b0752718-309a-41d3-80af-563483da9f5b","Type":"ContainerStarted","Data":"3fed4e056dea9f9734baee52079e17ba9ee932a92b3b83d588cb7d8c82ae230e"} Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.711722 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.711705825 podStartE2EDuration="2.711705825s" podCreationTimestamp="2025-12-03 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:55.705867349 +0000 UTC m=+1566.452199814" watchObservedRunningTime="2025-12-03 14:12:55.711705825 +0000 UTC m=+1566.458038280" Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.727517 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.727482817 podStartE2EDuration="2.727482817s" podCreationTimestamp="2025-12-03 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:55.723222272 +0000 UTC m=+1566.469554737" watchObservedRunningTime="2025-12-03 14:12:55.727482817 +0000 UTC m=+1566.473815272" Dec 03 14:12:55 crc kubenswrapper[4677]: I1203 14:12:55.746211 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.746190976 podStartE2EDuration="2.746190976s" podCreationTimestamp="2025-12-03 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:55.741005827 +0000 UTC m=+1566.487338292" watchObservedRunningTime="2025-12-03 14:12:55.746190976 +0000 UTC m=+1566.492523431" Dec 03 14:12:59 crc kubenswrapper[4677]: I1203 14:12:59.127995 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 14:12:59 crc kubenswrapper[4677]: I1203 14:12:59.220052 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 14:13:00 crc kubenswrapper[4677]: I1203 14:13:00.300538 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:13:00 crc kubenswrapper[4677]: I1203 14:13:00.300592 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:13:01 crc kubenswrapper[4677]: I1203 14:13:01.314169 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:01 crc kubenswrapper[4677]: I1203 14:13:01.314193 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:04 crc kubenswrapper[4677]: I1203 14:13:04.220173 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 14:13:04 crc kubenswrapper[4677]: I1203 14:13:04.247094 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 14:13:04 crc kubenswrapper[4677]: I1203 14:13:04.430017 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:13:04 crc kubenswrapper[4677]: I1203 14:13:04.430086 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:13:04 crc kubenswrapper[4677]: I1203 14:13:04.827995 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 14:13:05 crc kubenswrapper[4677]: I1203 14:13:05.471256 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:05 crc kubenswrapper[4677]: I1203 14:13:05.512193 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:08 crc kubenswrapper[4677]: I1203 14:13:08.437066 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:13:08 crc kubenswrapper[4677]: I1203 14:13:08.438488 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:13:10 crc kubenswrapper[4677]: I1203 14:13:10.310724 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:13:10 crc kubenswrapper[4677]: I1203 14:13:10.313745 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:13:10 crc kubenswrapper[4677]: I1203 14:13:10.322638 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:13:10 crc kubenswrapper[4677]: I1203 14:13:10.883182 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:13:11 crc kubenswrapper[4677]: I1203 14:13:11.890749 4677 generic.go:334] "Generic (PLEG): container finished" podID="67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" containerID="bd4cf25b1ae6dbc951dabe951610ae35ea15903d469d83a681b551add1ff2a2f" exitCode=137 Dec 03 14:13:11 crc kubenswrapper[4677]: I1203 14:13:11.892477 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b","Type":"ContainerDied","Data":"bd4cf25b1ae6dbc951dabe951610ae35ea15903d469d83a681b551add1ff2a2f"} Dec 03 14:13:11 crc kubenswrapper[4677]: I1203 14:13:11.892519 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b","Type":"ContainerDied","Data":"e81b0517ea90e3772c9d3a168a626ef7e54c8edb1c5df7f4d4e1a4da5868e3bf"} Dec 03 14:13:11 crc kubenswrapper[4677]: I1203 14:13:11.892534 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e81b0517ea90e3772c9d3a168a626ef7e54c8edb1c5df7f4d4e1a4da5868e3bf" Dec 03 14:13:11 crc kubenswrapper[4677]: I1203 14:13:11.983919 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.083588 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-config-data\") pod \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.083742 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z72hj\" (UniqueName: \"kubernetes.io/projected/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-kube-api-access-z72hj\") pod \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.083834 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-combined-ca-bundle\") pod \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\" (UID: \"67d2e8e2-6c98-4c30-9c6f-0817c401ed0b\") " Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.091461 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-kube-api-access-z72hj" (OuterVolumeSpecName: "kube-api-access-z72hj") pod "67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" (UID: "67d2e8e2-6c98-4c30-9c6f-0817c401ed0b"). InnerVolumeSpecName "kube-api-access-z72hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.116044 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" (UID: "67d2e8e2-6c98-4c30-9c6f-0817c401ed0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.117337 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-config-data" (OuterVolumeSpecName: "config-data") pod "67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" (UID: "67d2e8e2-6c98-4c30-9c6f-0817c401ed0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.186087 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.186118 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z72hj\" (UniqueName: \"kubernetes.io/projected/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-kube-api-access-z72hj\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.186129 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.903127 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.966198 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.982001 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.993581 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:13:12 crc kubenswrapper[4677]: E1203 14:13:12.994552 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.994575 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.994804 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 14:13:12 crc kubenswrapper[4677]: I1203 14:13:12.995621 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:12.998932 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:12.999148 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:12.999291 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.001606 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.103860 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.104027 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.104055 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.104086 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j26n8\" (UniqueName: \"kubernetes.io/projected/57acac06-fa90-4fa1-bf79-f23abca7645a-kube-api-access-j26n8\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.104178 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.207845 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.207979 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.208011 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.208056 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j26n8\" (UniqueName: \"kubernetes.io/projected/57acac06-fa90-4fa1-bf79-f23abca7645a-kube-api-access-j26n8\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.208142 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.212221 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.220473 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.220697 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.220842 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57acac06-fa90-4fa1-bf79-f23abca7645a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.227230 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j26n8\" (UniqueName: \"kubernetes.io/projected/57acac06-fa90-4fa1-bf79-f23abca7645a-kube-api-access-j26n8\") pod \"nova-cell1-novncproxy-0\" (UID: \"57acac06-fa90-4fa1-bf79-f23abca7645a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.318357 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.767440 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.912673 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"57acac06-fa90-4fa1-bf79-f23abca7645a","Type":"ContainerStarted","Data":"27ec91a6db34575024cc89aa101f72bf511ad423869a59d3375fedb10c95259c"} Dec 03 14:13:13 crc kubenswrapper[4677]: I1203 14:13:13.987096 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d2e8e2-6c98-4c30-9c6f-0817c401ed0b" path="/var/lib/kubelet/pods/67d2e8e2-6c98-4c30-9c6f-0817c401ed0b/volumes" Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.436868 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.437917 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.443726 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.447120 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.924172 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"57acac06-fa90-4fa1-bf79-f23abca7645a","Type":"ContainerStarted","Data":"5d2246a85789fad24e1633ef5e1afd2d22ce6f4bc5f0548120cd56527ac9ebe4"} Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.924253 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.936489 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:13:14 crc kubenswrapper[4677]: I1203 14:13:14.950571 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.950544912 podStartE2EDuration="2.950544912s" podCreationTimestamp="2025-12-03 14:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:13:14.943273528 +0000 UTC m=+1585.689606003" watchObservedRunningTime="2025-12-03 14:13:14.950544912 +0000 UTC m=+1585.696877377" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.174605 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-846d784fbc-wmfdq"] Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.176566 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.209574 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-846d784fbc-wmfdq"] Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.359767 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-svc\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.359853 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46ntn\" (UniqueName: \"kubernetes.io/projected/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-kube-api-access-46ntn\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.359909 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-swift-storage-0\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.360113 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-nb\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.360300 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-config\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.360466 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-sb\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.461979 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-svc\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.462036 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46ntn\" (UniqueName: \"kubernetes.io/projected/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-kube-api-access-46ntn\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.462069 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-swift-storage-0\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.462110 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-nb\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.462158 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-config\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.462210 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-sb\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.463227 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-swift-storage-0\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.463305 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-svc\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.463331 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-sb\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.463305 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-nb\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.463801 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-config\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.486834 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46ntn\" (UniqueName: \"kubernetes.io/projected/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-kube-api-access-46ntn\") pod \"dnsmasq-dns-846d784fbc-wmfdq\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:15 crc kubenswrapper[4677]: I1203 14:13:15.509934 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:16 crc kubenswrapper[4677]: I1203 14:13:16.011043 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-846d784fbc-wmfdq"] Dec 03 14:13:16 crc kubenswrapper[4677]: I1203 14:13:16.945176 4677 generic.go:334] "Generic (PLEG): container finished" podID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerID="2a95e17880c372c0bc5379a81c445f983175075784281af1c5de4f1f31901ad6" exitCode=0 Dec 03 14:13:16 crc kubenswrapper[4677]: I1203 14:13:16.947028 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" event={"ID":"7b69f0ff-660e-4535-8d71-10b8d67dfb2b","Type":"ContainerDied","Data":"2a95e17880c372c0bc5379a81c445f983175075784281af1c5de4f1f31901ad6"} Dec 03 14:13:16 crc kubenswrapper[4677]: I1203 14:13:16.947110 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" event={"ID":"7b69f0ff-660e-4535-8d71-10b8d67dfb2b","Type":"ContainerStarted","Data":"decaee75b304bc99129e0c7d31073862879e946d0bc3556d3c6445ec8df5d3a3"} Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.691694 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.793485 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.794015 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-central-agent" containerID="cri-o://b91a5deb92cad98bcb2c8b1bb662cde4d194cf37b5175dc43b5133efd8c2a20c" gracePeriod=30 Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.794156 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-notification-agent" containerID="cri-o://c30f57bfb3109e0309985c74f4a034508e57c525ec9a66c240862986887d6ded" gracePeriod=30 Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.794198 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="proxy-httpd" containerID="cri-o://8730132cc6488e77198dfa36fa1618e12ece424d3757ff62a0668cbcd1f63037" gracePeriod=30 Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.794456 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="sg-core" containerID="cri-o://5331a97bb1a0493c45473f17b93463b4d3a106cc4efdaff51a3e290f7c2585be" gracePeriod=30 Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.800056 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.201:3000/\": EOF" Dec 03 14:13:17 crc kubenswrapper[4677]: I1203 14:13:17.998806 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gsw7z"] Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.001565 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.006040 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" event={"ID":"7b69f0ff-660e-4535-8d71-10b8d67dfb2b","Type":"ContainerStarted","Data":"02f7b8f50c4afc59b154aee5813eecbde9a5e88ee91aa969c37d05cf52d1c40e"} Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.007087 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.014314 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsw7z"] Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.021324 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lrh9\" (UniqueName: \"kubernetes.io/projected/1ced0c0f-42d8-4695-8f16-310ef7ab0046-kube-api-access-8lrh9\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.021436 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-catalog-content\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.021572 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-utilities\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.045206 4677 generic.go:334] "Generic (PLEG): container finished" podID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerID="5331a97bb1a0493c45473f17b93463b4d3a106cc4efdaff51a3e290f7c2585be" exitCode=2 Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.045308 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerDied","Data":"5331a97bb1a0493c45473f17b93463b4d3a106cc4efdaff51a3e290f7c2585be"} Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.045485 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-log" containerID="cri-o://7e79b4c0cb29b438e3276028cc2551fa2016451bbc72028e590fa0f2d6985e09" gracePeriod=30 Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.045559 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-api" containerID="cri-o://7a2fdf98a127eae0b332d0b76e17baa09257f3b5ab3cfe89e6ab4f6a9dd08810" gracePeriod=30 Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.122922 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lrh9\" (UniqueName: \"kubernetes.io/projected/1ced0c0f-42d8-4695-8f16-310ef7ab0046-kube-api-access-8lrh9\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.122982 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-catalog-content\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.123178 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-utilities\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.123630 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-utilities\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.125167 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-catalog-content\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.144773 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lrh9\" (UniqueName: \"kubernetes.io/projected/1ced0c0f-42d8-4695-8f16-310ef7ab0046-kube-api-access-8lrh9\") pod \"certified-operators-gsw7z\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.319357 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.408030 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:18 crc kubenswrapper[4677]: W1203 14:13:18.930497 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ced0c0f_42d8_4695_8f16_310ef7ab0046.slice/crio-1914df0917a724046529fdc86840fea8ca7b20ebd8ecc78bcb55ec476fc562e9 WatchSource:0}: Error finding container 1914df0917a724046529fdc86840fea8ca7b20ebd8ecc78bcb55ec476fc562e9: Status 404 returned error can't find the container with id 1914df0917a724046529fdc86840fea8ca7b20ebd8ecc78bcb55ec476fc562e9 Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.933441 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" podStartSLOduration=3.933421279 podStartE2EDuration="3.933421279s" podCreationTimestamp="2025-12-03 14:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:13:18.086856529 +0000 UTC m=+1588.833188994" watchObservedRunningTime="2025-12-03 14:13:18.933421279 +0000 UTC m=+1589.679753744" Dec 03 14:13:18 crc kubenswrapper[4677]: I1203 14:13:18.934474 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsw7z"] Dec 03 14:13:19 crc kubenswrapper[4677]: I1203 14:13:19.129460 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0752718-309a-41d3-80af-563483da9f5b" containerID="7e79b4c0cb29b438e3276028cc2551fa2016451bbc72028e590fa0f2d6985e09" exitCode=143 Dec 03 14:13:19 crc kubenswrapper[4677]: I1203 14:13:19.129797 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b0752718-309a-41d3-80af-563483da9f5b","Type":"ContainerDied","Data":"7e79b4c0cb29b438e3276028cc2551fa2016451bbc72028e590fa0f2d6985e09"} Dec 03 14:13:19 crc kubenswrapper[4677]: I1203 14:13:19.138341 4677 generic.go:334] "Generic (PLEG): container finished" podID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerID="8730132cc6488e77198dfa36fa1618e12ece424d3757ff62a0668cbcd1f63037" exitCode=0 Dec 03 14:13:19 crc kubenswrapper[4677]: I1203 14:13:19.138388 4677 generic.go:334] "Generic (PLEG): container finished" podID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerID="b91a5deb92cad98bcb2c8b1bb662cde4d194cf37b5175dc43b5133efd8c2a20c" exitCode=0 Dec 03 14:13:19 crc kubenswrapper[4677]: I1203 14:13:19.138446 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerDied","Data":"8730132cc6488e77198dfa36fa1618e12ece424d3757ff62a0668cbcd1f63037"} Dec 03 14:13:19 crc kubenswrapper[4677]: I1203 14:13:19.138474 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerDied","Data":"b91a5deb92cad98bcb2c8b1bb662cde4d194cf37b5175dc43b5133efd8c2a20c"} Dec 03 14:13:19 crc kubenswrapper[4677]: I1203 14:13:19.142012 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsw7z" event={"ID":"1ced0c0f-42d8-4695-8f16-310ef7ab0046","Type":"ContainerStarted","Data":"1914df0917a724046529fdc86840fea8ca7b20ebd8ecc78bcb55ec476fc562e9"} Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.151714 4677 generic.go:334] "Generic (PLEG): container finished" podID="b0752718-309a-41d3-80af-563483da9f5b" containerID="7a2fdf98a127eae0b332d0b76e17baa09257f3b5ab3cfe89e6ab4f6a9dd08810" exitCode=0 Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.151792 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b0752718-309a-41d3-80af-563483da9f5b","Type":"ContainerDied","Data":"7a2fdf98a127eae0b332d0b76e17baa09257f3b5ab3cfe89e6ab4f6a9dd08810"} Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.154172 4677 generic.go:334] "Generic (PLEG): container finished" podID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerID="58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8" exitCode=0 Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.154239 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsw7z" event={"ID":"1ced0c0f-42d8-4695-8f16-310ef7ab0046","Type":"ContainerDied","Data":"58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8"} Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.367341 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.201:3000/\": dial tcp 10.217.0.201:3000: connect: connection refused" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.462055 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.584482 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-combined-ca-bundle\") pod \"b0752718-309a-41d3-80af-563483da9f5b\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.584625 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0752718-309a-41d3-80af-563483da9f5b-logs\") pod \"b0752718-309a-41d3-80af-563483da9f5b\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.584654 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs7w9\" (UniqueName: \"kubernetes.io/projected/b0752718-309a-41d3-80af-563483da9f5b-kube-api-access-rs7w9\") pod \"b0752718-309a-41d3-80af-563483da9f5b\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.584705 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-config-data\") pod \"b0752718-309a-41d3-80af-563483da9f5b\" (UID: \"b0752718-309a-41d3-80af-563483da9f5b\") " Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.585065 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0752718-309a-41d3-80af-563483da9f5b-logs" (OuterVolumeSpecName: "logs") pod "b0752718-309a-41d3-80af-563483da9f5b" (UID: "b0752718-309a-41d3-80af-563483da9f5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.585241 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0752718-309a-41d3-80af-563483da9f5b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.592575 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0752718-309a-41d3-80af-563483da9f5b-kube-api-access-rs7w9" (OuterVolumeSpecName: "kube-api-access-rs7w9") pod "b0752718-309a-41d3-80af-563483da9f5b" (UID: "b0752718-309a-41d3-80af-563483da9f5b"). InnerVolumeSpecName "kube-api-access-rs7w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.616258 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-config-data" (OuterVolumeSpecName: "config-data") pod "b0752718-309a-41d3-80af-563483da9f5b" (UID: "b0752718-309a-41d3-80af-563483da9f5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.656772 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0752718-309a-41d3-80af-563483da9f5b" (UID: "b0752718-309a-41d3-80af-563483da9f5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.686921 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.687203 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs7w9\" (UniqueName: \"kubernetes.io/projected/b0752718-309a-41d3-80af-563483da9f5b-kube-api-access-rs7w9\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:20 crc kubenswrapper[4677]: I1203 14:13:20.687268 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0752718-309a-41d3-80af-563483da9f5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.168363 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b0752718-309a-41d3-80af-563483da9f5b","Type":"ContainerDied","Data":"3fed4e056dea9f9734baee52079e17ba9ee932a92b3b83d588cb7d8c82ae230e"} Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.168707 4677 scope.go:117] "RemoveContainer" containerID="7a2fdf98a127eae0b332d0b76e17baa09257f3b5ab3cfe89e6ab4f6a9dd08810" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.168455 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.205143 4677 scope.go:117] "RemoveContainer" containerID="7e79b4c0cb29b438e3276028cc2551fa2016451bbc72028e590fa0f2d6985e09" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.216367 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.237523 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.255090 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:21 crc kubenswrapper[4677]: E1203 14:13:21.255652 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-api" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.255672 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-api" Dec 03 14:13:21 crc kubenswrapper[4677]: E1203 14:13:21.255710 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-log" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.255717 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-log" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.265307 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-api" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.265372 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0752718-309a-41d3-80af-563483da9f5b" containerName="nova-api-log" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.266778 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.266879 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.269095 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.271146 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.271363 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.410040 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-config-data\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.410118 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-public-tls-certs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.410143 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.410163 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.410179 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74cj9\" (UniqueName: \"kubernetes.io/projected/22e11417-f484-48b1-a9c6-87b77c7470eb-kube-api-access-74cj9\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.410196 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22e11417-f484-48b1-a9c6-87b77c7470eb-logs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.512109 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-public-tls-certs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.512160 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.512190 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.512213 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74cj9\" (UniqueName: \"kubernetes.io/projected/22e11417-f484-48b1-a9c6-87b77c7470eb-kube-api-access-74cj9\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.512237 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22e11417-f484-48b1-a9c6-87b77c7470eb-logs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.512409 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-config-data\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.513462 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22e11417-f484-48b1-a9c6-87b77c7470eb-logs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.518448 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.518568 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-config-data\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.518604 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.519064 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-public-tls-certs\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.533631 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74cj9\" (UniqueName: \"kubernetes.io/projected/22e11417-f484-48b1-a9c6-87b77c7470eb-kube-api-access-74cj9\") pod \"nova-api-0\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.597757 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:21 crc kubenswrapper[4677]: I1203 14:13:21.991011 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0752718-309a-41d3-80af-563483da9f5b" path="/var/lib/kubelet/pods/b0752718-309a-41d3-80af-563483da9f5b/volumes" Dec 03 14:13:22 crc kubenswrapper[4677]: I1203 14:13:22.133048 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:22 crc kubenswrapper[4677]: W1203 14:13:22.138097 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22e11417_f484_48b1_a9c6_87b77c7470eb.slice/crio-ebc2a0ea9004aa68d04d721d61d6505a47fab0ee312dc68879b7c3f04070a90f WatchSource:0}: Error finding container ebc2a0ea9004aa68d04d721d61d6505a47fab0ee312dc68879b7c3f04070a90f: Status 404 returned error can't find the container with id ebc2a0ea9004aa68d04d721d61d6505a47fab0ee312dc68879b7c3f04070a90f Dec 03 14:13:22 crc kubenswrapper[4677]: I1203 14:13:22.187869 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsw7z" event={"ID":"1ced0c0f-42d8-4695-8f16-310ef7ab0046","Type":"ContainerStarted","Data":"9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f"} Dec 03 14:13:22 crc kubenswrapper[4677]: I1203 14:13:22.198211 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22e11417-f484-48b1-a9c6-87b77c7470eb","Type":"ContainerStarted","Data":"ebc2a0ea9004aa68d04d721d61d6505a47fab0ee312dc68879b7c3f04070a90f"} Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.224708 4677 generic.go:334] "Generic (PLEG): container finished" podID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerID="c30f57bfb3109e0309985c74f4a034508e57c525ec9a66c240862986887d6ded" exitCode=0 Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.225445 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerDied","Data":"c30f57bfb3109e0309985c74f4a034508e57c525ec9a66c240862986887d6ded"} Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.229080 4677 generic.go:334] "Generic (PLEG): container finished" podID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerID="9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f" exitCode=0 Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.229136 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsw7z" event={"ID":"1ced0c0f-42d8-4695-8f16-310ef7ab0046","Type":"ContainerDied","Data":"9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f"} Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.233598 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22e11417-f484-48b1-a9c6-87b77c7470eb","Type":"ContainerStarted","Data":"9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11"} Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.233620 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22e11417-f484-48b1-a9c6-87b77c7470eb","Type":"ContainerStarted","Data":"f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab"} Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.310132 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.3101052429999998 podStartE2EDuration="2.310105243s" podCreationTimestamp="2025-12-03 14:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:13:23.289479111 +0000 UTC m=+1594.035811576" watchObservedRunningTime="2025-12-03 14:13:23.310105243 +0000 UTC m=+1594.056437718" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.318858 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.343968 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.469133 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.558252 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkjmd\" (UniqueName: \"kubernetes.io/projected/c0b61325-6e17-4a89-a83a-0f1de7d2db43-kube-api-access-jkjmd\") pod \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.558345 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-scripts\") pod \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.558437 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-run-httpd\") pod \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.558499 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-combined-ca-bundle\") pod \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.558530 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-log-httpd\") pod \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.558553 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-config-data\") pod \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.558649 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-sg-core-conf-yaml\") pod \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\" (UID: \"c0b61325-6e17-4a89-a83a-0f1de7d2db43\") " Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.559247 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c0b61325-6e17-4a89-a83a-0f1de7d2db43" (UID: "c0b61325-6e17-4a89-a83a-0f1de7d2db43"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.559609 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c0b61325-6e17-4a89-a83a-0f1de7d2db43" (UID: "c0b61325-6e17-4a89-a83a-0f1de7d2db43"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.563731 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-scripts" (OuterVolumeSpecName: "scripts") pod "c0b61325-6e17-4a89-a83a-0f1de7d2db43" (UID: "c0b61325-6e17-4a89-a83a-0f1de7d2db43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.577193 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0b61325-6e17-4a89-a83a-0f1de7d2db43-kube-api-access-jkjmd" (OuterVolumeSpecName: "kube-api-access-jkjmd") pod "c0b61325-6e17-4a89-a83a-0f1de7d2db43" (UID: "c0b61325-6e17-4a89-a83a-0f1de7d2db43"). InnerVolumeSpecName "kube-api-access-jkjmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.590630 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c0b61325-6e17-4a89-a83a-0f1de7d2db43" (UID: "c0b61325-6e17-4a89-a83a-0f1de7d2db43"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.650873 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0b61325-6e17-4a89-a83a-0f1de7d2db43" (UID: "c0b61325-6e17-4a89-a83a-0f1de7d2db43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.662215 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.662246 4677 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.662255 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.662264 4677 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0b61325-6e17-4a89-a83a-0f1de7d2db43-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.662273 4677 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.662282 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkjmd\" (UniqueName: \"kubernetes.io/projected/c0b61325-6e17-4a89-a83a-0f1de7d2db43-kube-api-access-jkjmd\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.668484 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-config-data" (OuterVolumeSpecName: "config-data") pod "c0b61325-6e17-4a89-a83a-0f1de7d2db43" (UID: "c0b61325-6e17-4a89-a83a-0f1de7d2db43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:23 crc kubenswrapper[4677]: I1203 14:13:23.763691 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b61325-6e17-4a89-a83a-0f1de7d2db43-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.246779 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0b61325-6e17-4a89-a83a-0f1de7d2db43","Type":"ContainerDied","Data":"259e232ac64fa65069d8f9ea390ebf0f47bbd9bb43bd76cf9df41539366601e4"} Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.246964 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.247069 4677 scope.go:117] "RemoveContainer" containerID="8730132cc6488e77198dfa36fa1618e12ece424d3757ff62a0668cbcd1f63037" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.264890 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.310369 4677 scope.go:117] "RemoveContainer" containerID="5331a97bb1a0493c45473f17b93463b4d3a106cc4efdaff51a3e290f7c2585be" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.315627 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.352301 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.357324 4677 scope.go:117] "RemoveContainer" containerID="c30f57bfb3109e0309985c74f4a034508e57c525ec9a66c240862986887d6ded" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.372650 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:13:24 crc kubenswrapper[4677]: E1203 14:13:24.374018 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-central-agent" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374045 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-central-agent" Dec 03 14:13:24 crc kubenswrapper[4677]: E1203 14:13:24.374106 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="proxy-httpd" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374116 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="proxy-httpd" Dec 03 14:13:24 crc kubenswrapper[4677]: E1203 14:13:24.374136 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="sg-core" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374144 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="sg-core" Dec 03 14:13:24 crc kubenswrapper[4677]: E1203 14:13:24.374158 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-notification-agent" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374167 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-notification-agent" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374436 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="proxy-httpd" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374462 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="sg-core" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374488 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-notification-agent" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.374505 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" containerName="ceilometer-central-agent" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.377227 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.379715 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.381873 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.386658 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.393195 4677 scope.go:117] "RemoveContainer" containerID="b91a5deb92cad98bcb2c8b1bb662cde4d194cf37b5175dc43b5133efd8c2a20c" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.483080 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm8ql\" (UniqueName: \"kubernetes.io/projected/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-kube-api-access-mm8ql\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.483127 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-scripts\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.483146 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-run-httpd\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.483163 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-config-data\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.483483 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.483535 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.483912 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-log-httpd\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.553016 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-wxzts"] Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.555808 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.558179 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.558463 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.566849 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-wxzts"] Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.591130 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-log-httpd\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.591223 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm8ql\" (UniqueName: \"kubernetes.io/projected/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-kube-api-access-mm8ql\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.591249 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-scripts\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.591262 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-run-httpd\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.591278 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-config-data\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.591383 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.591399 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.592286 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-log-httpd\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.592545 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-run-httpd\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.597848 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-config-data\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.598521 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-scripts\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.599613 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.610323 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm8ql\" (UniqueName: \"kubernetes.io/projected/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-kube-api-access-mm8ql\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.619273 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.693440 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-config-data\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.693835 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.694060 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-scripts\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.694093 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmqqx\" (UniqueName: \"kubernetes.io/projected/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-kube-api-access-bmqqx\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.704087 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.795987 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-scripts\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.796030 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmqqx\" (UniqueName: \"kubernetes.io/projected/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-kube-api-access-bmqqx\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.796112 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-config-data\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.796151 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.803519 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.803923 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-scripts\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.803941 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-config-data\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.819664 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmqqx\" (UniqueName: \"kubernetes.io/projected/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-kube-api-access-bmqqx\") pod \"nova-cell1-cell-mapping-wxzts\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:24 crc kubenswrapper[4677]: I1203 14:13:24.888587 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.179619 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.258863 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerStarted","Data":"f6b6669c0e56322d1df2426aadb317d06ae848871239c3e6d039473c7db15c84"} Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.263469 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsw7z" event={"ID":"1ced0c0f-42d8-4695-8f16-310ef7ab0046","Type":"ContainerStarted","Data":"2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042"} Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.280200 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gsw7z" podStartSLOduration=4.334978925 podStartE2EDuration="8.280179395s" podCreationTimestamp="2025-12-03 14:13:17 +0000 UTC" firstStartedPulling="2025-12-03 14:13:20.155807775 +0000 UTC m=+1590.902140230" lastFinishedPulling="2025-12-03 14:13:24.101008225 +0000 UTC m=+1594.847340700" observedRunningTime="2025-12-03 14:13:25.277681088 +0000 UTC m=+1596.024013553" watchObservedRunningTime="2025-12-03 14:13:25.280179395 +0000 UTC m=+1596.026511860" Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.401811 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-wxzts"] Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.511217 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.584652 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86755f6c8c-s8sbl"] Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.584920 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" podUID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerName="dnsmasq-dns" containerID="cri-o://3d58c3e477a90f3bd83a928441860da1400291befd46e8aba7c4d491347e6e6f" gracePeriod=10 Dec 03 14:13:25 crc kubenswrapper[4677]: I1203 14:13:25.991275 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0b61325-6e17-4a89-a83a-0f1de7d2db43" path="/var/lib/kubelet/pods/c0b61325-6e17-4a89-a83a-0f1de7d2db43/volumes" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.282854 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerStarted","Data":"effea24d757bba312a7308f5cae2de09fead0cec9d7e44bf0d5e9576bd05c038"} Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.297048 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerID="3d58c3e477a90f3bd83a928441860da1400291befd46e8aba7c4d491347e6e6f" exitCode=0 Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.297155 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" event={"ID":"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf","Type":"ContainerDied","Data":"3d58c3e477a90f3bd83a928441860da1400291befd46e8aba7c4d491347e6e6f"} Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.314664 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wxzts" event={"ID":"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf","Type":"ContainerStarted","Data":"85df570ee32a8fd9b37707d8d51e419d2f0db38d23b67a8e9dc513992a8ea6cc"} Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.314746 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wxzts" event={"ID":"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf","Type":"ContainerStarted","Data":"c3495aa9df09297cf48dec75b368520ee33d18cfe2776d0acc09221493aa8de8"} Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.342398 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-wxzts" podStartSLOduration=2.342375759 podStartE2EDuration="2.342375759s" podCreationTimestamp="2025-12-03 14:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:13:26.33234003 +0000 UTC m=+1597.078672485" watchObservedRunningTime="2025-12-03 14:13:26.342375759 +0000 UTC m=+1597.088708224" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.712295 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.841991 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-config\") pod \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.842111 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-svc\") pod \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.842227 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-swift-storage-0\") pod \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.842247 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-nb\") pod \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.842309 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-sb\") pod \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.842350 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5h8g\" (UniqueName: \"kubernetes.io/projected/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-kube-api-access-r5h8g\") pod \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\" (UID: \"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf\") " Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.850146 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-kube-api-access-r5h8g" (OuterVolumeSpecName: "kube-api-access-r5h8g") pod "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" (UID: "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf"). InnerVolumeSpecName "kube-api-access-r5h8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.903866 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" (UID: "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.907484 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" (UID: "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.909936 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-config" (OuterVolumeSpecName: "config") pod "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" (UID: "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.915510 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" (UID: "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.921026 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" (UID: "a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.944942 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.944993 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.945005 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.945014 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.945022 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5h8g\" (UniqueName: \"kubernetes.io/projected/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-kube-api-access-r5h8g\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:26 crc kubenswrapper[4677]: I1203 14:13:26.945032 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.348479 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" event={"ID":"a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf","Type":"ContainerDied","Data":"bacc73612c7259bf1b472d4107007591a25c8bf1f9fe6119afe04034402219be"} Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.348500 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86755f6c8c-s8sbl" Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.348849 4677 scope.go:117] "RemoveContainer" containerID="3d58c3e477a90f3bd83a928441860da1400291befd46e8aba7c4d491347e6e6f" Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.353967 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerStarted","Data":"73aab7ebcfc7c09953d7083f2578edcbd60ec1347f1a08b6d6970e665f5ddc17"} Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.354018 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerStarted","Data":"7dfe5ffdcda37accb79174c1e77f2afbc04253b47ca875e87e5d58eaa6a6132f"} Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.380402 4677 scope.go:117] "RemoveContainer" containerID="e2eb609ad114f74e2cf57387fbc15bf916dac81b877cd1a78780261721ebca68" Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.391501 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86755f6c8c-s8sbl"] Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.400719 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86755f6c8c-s8sbl"] Dec 03 14:13:27 crc kubenswrapper[4677]: I1203 14:13:27.989560 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" path="/var/lib/kubelet/pods/a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf/volumes" Dec 03 14:13:28 crc kubenswrapper[4677]: I1203 14:13:28.408550 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:28 crc kubenswrapper[4677]: I1203 14:13:28.408887 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:28 crc kubenswrapper[4677]: I1203 14:13:28.476856 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:29 crc kubenswrapper[4677]: I1203 14:13:29.385679 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerStarted","Data":"9e6b5b8a16aa085236cb2a602bdfc3b1d473ab688a6b259b5fef010f6733f1a1"} Dec 03 14:13:29 crc kubenswrapper[4677]: I1203 14:13:29.386064 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:13:29 crc kubenswrapper[4677]: I1203 14:13:29.413230 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.060887456 podStartE2EDuration="5.413208426s" podCreationTimestamp="2025-12-03 14:13:24 +0000 UTC" firstStartedPulling="2025-12-03 14:13:25.17857228 +0000 UTC m=+1595.924904735" lastFinishedPulling="2025-12-03 14:13:28.53089324 +0000 UTC m=+1599.277225705" observedRunningTime="2025-12-03 14:13:29.408963343 +0000 UTC m=+1600.155295798" watchObservedRunningTime="2025-12-03 14:13:29.413208426 +0000 UTC m=+1600.159540891" Dec 03 14:13:29 crc kubenswrapper[4677]: I1203 14:13:29.439078 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:29 crc kubenswrapper[4677]: I1203 14:13:29.488773 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gsw7z"] Dec 03 14:13:31 crc kubenswrapper[4677]: I1203 14:13:31.406183 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gsw7z" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="registry-server" containerID="cri-o://2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042" gracePeriod=2 Dec 03 14:13:31 crc kubenswrapper[4677]: I1203 14:13:31.600594 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:13:31 crc kubenswrapper[4677]: I1203 14:13:31.600932 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:13:31 crc kubenswrapper[4677]: I1203 14:13:31.988642 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.153320 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-utilities\") pod \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.153441 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-catalog-content\") pod \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.153544 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lrh9\" (UniqueName: \"kubernetes.io/projected/1ced0c0f-42d8-4695-8f16-310ef7ab0046-kube-api-access-8lrh9\") pod \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\" (UID: \"1ced0c0f-42d8-4695-8f16-310ef7ab0046\") " Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.155351 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-utilities" (OuterVolumeSpecName: "utilities") pod "1ced0c0f-42d8-4695-8f16-310ef7ab0046" (UID: "1ced0c0f-42d8-4695-8f16-310ef7ab0046"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.190476 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ced0c0f-42d8-4695-8f16-310ef7ab0046-kube-api-access-8lrh9" (OuterVolumeSpecName: "kube-api-access-8lrh9") pod "1ced0c0f-42d8-4695-8f16-310ef7ab0046" (UID: "1ced0c0f-42d8-4695-8f16-310ef7ab0046"). InnerVolumeSpecName "kube-api-access-8lrh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.205869 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ced0c0f-42d8-4695-8f16-310ef7ab0046" (UID: "1ced0c0f-42d8-4695-8f16-310ef7ab0046"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.256292 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.256329 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ced0c0f-42d8-4695-8f16-310ef7ab0046-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.256342 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lrh9\" (UniqueName: \"kubernetes.io/projected/1ced0c0f-42d8-4695-8f16-310ef7ab0046-kube-api-access-8lrh9\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.418468 4677 generic.go:334] "Generic (PLEG): container finished" podID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerID="2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042" exitCode=0 Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.418535 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsw7z" event={"ID":"1ced0c0f-42d8-4695-8f16-310ef7ab0046","Type":"ContainerDied","Data":"2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042"} Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.418617 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsw7z" event={"ID":"1ced0c0f-42d8-4695-8f16-310ef7ab0046","Type":"ContainerDied","Data":"1914df0917a724046529fdc86840fea8ca7b20ebd8ecc78bcb55ec476fc562e9"} Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.418547 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsw7z" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.418637 4677 scope.go:117] "RemoveContainer" containerID="2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.421143 4677 generic.go:334] "Generic (PLEG): container finished" podID="ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" containerID="85df570ee32a8fd9b37707d8d51e419d2f0db38d23b67a8e9dc513992a8ea6cc" exitCode=0 Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.421198 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wxzts" event={"ID":"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf","Type":"ContainerDied","Data":"85df570ee32a8fd9b37707d8d51e419d2f0db38d23b67a8e9dc513992a8ea6cc"} Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.462341 4677 scope.go:117] "RemoveContainer" containerID="9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.464504 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gsw7z"] Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.476265 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gsw7z"] Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.491114 4677 scope.go:117] "RemoveContainer" containerID="58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.532674 4677 scope.go:117] "RemoveContainer" containerID="2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042" Dec 03 14:13:32 crc kubenswrapper[4677]: E1203 14:13:32.533486 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042\": container with ID starting with 2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042 not found: ID does not exist" containerID="2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.533546 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042"} err="failed to get container status \"2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042\": rpc error: code = NotFound desc = could not find container \"2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042\": container with ID starting with 2b003230302c7e0eea6e53cbe640a06f2d806da22d095681723c1d8171bf6042 not found: ID does not exist" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.533587 4677 scope.go:117] "RemoveContainer" containerID="9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f" Dec 03 14:13:32 crc kubenswrapper[4677]: E1203 14:13:32.535587 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f\": container with ID starting with 9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f not found: ID does not exist" containerID="9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.535635 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f"} err="failed to get container status \"9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f\": rpc error: code = NotFound desc = could not find container \"9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f\": container with ID starting with 9c47fdba5ce80a9cd9933a040dd6d8b79254df73906287bc38bf0a44169cfb7f not found: ID does not exist" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.535658 4677 scope.go:117] "RemoveContainer" containerID="58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8" Dec 03 14:13:32 crc kubenswrapper[4677]: E1203 14:13:32.536156 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8\": container with ID starting with 58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8 not found: ID does not exist" containerID="58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.536213 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8"} err="failed to get container status \"58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8\": rpc error: code = NotFound desc = could not find container \"58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8\": container with ID starting with 58453cb98808a24a948fd9f0aee67cb51427262deff86c4110c948b22f9f00d8 not found: ID does not exist" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.654363 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.218:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:32 crc kubenswrapper[4677]: I1203 14:13:32.654905 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.218:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.852387 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.987017 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" path="/var/lib/kubelet/pods/1ced0c0f-42d8-4695-8f16-310ef7ab0046/volumes" Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.990636 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-combined-ca-bundle\") pod \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.990683 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-config-data\") pod \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.990778 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-scripts\") pod \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.990823 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmqqx\" (UniqueName: \"kubernetes.io/projected/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-kube-api-access-bmqqx\") pod \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\" (UID: \"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf\") " Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.996089 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-kube-api-access-bmqqx" (OuterVolumeSpecName: "kube-api-access-bmqqx") pod "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" (UID: "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf"). InnerVolumeSpecName "kube-api-access-bmqqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:33 crc kubenswrapper[4677]: I1203 14:13:33.996676 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-scripts" (OuterVolumeSpecName: "scripts") pod "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" (UID: "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.022999 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" (UID: "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.024857 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-config-data" (OuterVolumeSpecName: "config-data") pod "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" (UID: "ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.102979 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.103020 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmqqx\" (UniqueName: \"kubernetes.io/projected/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-kube-api-access-bmqqx\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.103032 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.103044 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.486443 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wxzts" event={"ID":"ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf","Type":"ContainerDied","Data":"c3495aa9df09297cf48dec75b368520ee33d18cfe2776d0acc09221493aa8de8"} Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.486787 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3495aa9df09297cf48dec75b368520ee33d18cfe2776d0acc09221493aa8de8" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.486851 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wxzts" Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.633170 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.633673 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8d042f25-044e-431e-b353-9f37399f41f3" containerName="nova-scheduler-scheduler" containerID="cri-o://9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de" gracePeriod=30 Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.668144 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.668389 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-log" containerID="cri-o://f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab" gracePeriod=30 Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.668800 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-api" containerID="cri-o://9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11" gracePeriod=30 Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.691864 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.692157 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-log" containerID="cri-o://95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e" gracePeriod=30 Dec 03 14:13:34 crc kubenswrapper[4677]: I1203 14:13:34.692226 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-metadata" containerID="cri-o://aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6" gracePeriod=30 Dec 03 14:13:35 crc kubenswrapper[4677]: I1203 14:13:35.497423 4677 generic.go:334] "Generic (PLEG): container finished" podID="7e74337e-3593-494a-8a23-1b787d670087" containerID="95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e" exitCode=143 Dec 03 14:13:35 crc kubenswrapper[4677]: I1203 14:13:35.497545 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e74337e-3593-494a-8a23-1b787d670087","Type":"ContainerDied","Data":"95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e"} Dec 03 14:13:35 crc kubenswrapper[4677]: I1203 14:13:35.499921 4677 generic.go:334] "Generic (PLEG): container finished" podID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerID="f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab" exitCode=143 Dec 03 14:13:35 crc kubenswrapper[4677]: I1203 14:13:35.499977 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22e11417-f484-48b1-a9c6-87b77c7470eb","Type":"ContainerDied","Data":"f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab"} Dec 03 14:13:35 crc kubenswrapper[4677]: I1203 14:13:35.554378 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": read tcp 10.217.0.2:47292->10.217.0.211:8775: read: connection reset by peer" Dec 03 14:13:35 crc kubenswrapper[4677]: I1203 14:13:35.554407 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": read tcp 10.217.0.2:47298->10.217.0.211:8775: read: connection reset by peer" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.002232 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.169589 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57mhb\" (UniqueName: \"kubernetes.io/projected/7e74337e-3593-494a-8a23-1b787d670087-kube-api-access-57mhb\") pod \"7e74337e-3593-494a-8a23-1b787d670087\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.169765 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-config-data\") pod \"7e74337e-3593-494a-8a23-1b787d670087\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.169802 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e74337e-3593-494a-8a23-1b787d670087-logs\") pod \"7e74337e-3593-494a-8a23-1b787d670087\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.169908 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-combined-ca-bundle\") pod \"7e74337e-3593-494a-8a23-1b787d670087\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.169976 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-nova-metadata-tls-certs\") pod \"7e74337e-3593-494a-8a23-1b787d670087\" (UID: \"7e74337e-3593-494a-8a23-1b787d670087\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.171290 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e74337e-3593-494a-8a23-1b787d670087-logs" (OuterVolumeSpecName: "logs") pod "7e74337e-3593-494a-8a23-1b787d670087" (UID: "7e74337e-3593-494a-8a23-1b787d670087"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.176705 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e74337e-3593-494a-8a23-1b787d670087-kube-api-access-57mhb" (OuterVolumeSpecName: "kube-api-access-57mhb") pod "7e74337e-3593-494a-8a23-1b787d670087" (UID: "7e74337e-3593-494a-8a23-1b787d670087"). InnerVolumeSpecName "kube-api-access-57mhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.222555 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-config-data" (OuterVolumeSpecName: "config-data") pod "7e74337e-3593-494a-8a23-1b787d670087" (UID: "7e74337e-3593-494a-8a23-1b787d670087"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.223888 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e74337e-3593-494a-8a23-1b787d670087" (UID: "7e74337e-3593-494a-8a23-1b787d670087"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.271968 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.272000 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57mhb\" (UniqueName: \"kubernetes.io/projected/7e74337e-3593-494a-8a23-1b787d670087-kube-api-access-57mhb\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.272010 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.272019 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e74337e-3593-494a-8a23-1b787d670087-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.275968 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7e74337e-3593-494a-8a23-1b787d670087" (UID: "7e74337e-3593-494a-8a23-1b787d670087"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.373850 4677 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e74337e-3593-494a-8a23-1b787d670087-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.387310 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.474598 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74cj9\" (UniqueName: \"kubernetes.io/projected/22e11417-f484-48b1-a9c6-87b77c7470eb-kube-api-access-74cj9\") pod \"22e11417-f484-48b1-a9c6-87b77c7470eb\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.474927 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-internal-tls-certs\") pod \"22e11417-f484-48b1-a9c6-87b77c7470eb\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.474981 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-config-data\") pod \"22e11417-f484-48b1-a9c6-87b77c7470eb\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.475138 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-public-tls-certs\") pod \"22e11417-f484-48b1-a9c6-87b77c7470eb\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.475178 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22e11417-f484-48b1-a9c6-87b77c7470eb-logs\") pod \"22e11417-f484-48b1-a9c6-87b77c7470eb\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.475204 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-combined-ca-bundle\") pod \"22e11417-f484-48b1-a9c6-87b77c7470eb\" (UID: \"22e11417-f484-48b1-a9c6-87b77c7470eb\") " Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.475871 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22e11417-f484-48b1-a9c6-87b77c7470eb-logs" (OuterVolumeSpecName: "logs") pod "22e11417-f484-48b1-a9c6-87b77c7470eb" (UID: "22e11417-f484-48b1-a9c6-87b77c7470eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.478774 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22e11417-f484-48b1-a9c6-87b77c7470eb-kube-api-access-74cj9" (OuterVolumeSpecName: "kube-api-access-74cj9") pod "22e11417-f484-48b1-a9c6-87b77c7470eb" (UID: "22e11417-f484-48b1-a9c6-87b77c7470eb"). InnerVolumeSpecName "kube-api-access-74cj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.513397 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22e11417-f484-48b1-a9c6-87b77c7470eb" (UID: "22e11417-f484-48b1-a9c6-87b77c7470eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.518802 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-config-data" (OuterVolumeSpecName: "config-data") pod "22e11417-f484-48b1-a9c6-87b77c7470eb" (UID: "22e11417-f484-48b1-a9c6-87b77c7470eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.520762 4677 generic.go:334] "Generic (PLEG): container finished" podID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerID="9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11" exitCode=0 Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.520858 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22e11417-f484-48b1-a9c6-87b77c7470eb","Type":"ContainerDied","Data":"9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11"} Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.520915 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22e11417-f484-48b1-a9c6-87b77c7470eb","Type":"ContainerDied","Data":"ebc2a0ea9004aa68d04d721d61d6505a47fab0ee312dc68879b7c3f04070a90f"} Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.520937 4677 scope.go:117] "RemoveContainer" containerID="9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.521192 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.525500 4677 generic.go:334] "Generic (PLEG): container finished" podID="7e74337e-3593-494a-8a23-1b787d670087" containerID="aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6" exitCode=0 Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.525539 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e74337e-3593-494a-8a23-1b787d670087","Type":"ContainerDied","Data":"aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6"} Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.525567 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e74337e-3593-494a-8a23-1b787d670087","Type":"ContainerDied","Data":"194347dc3e32c35785a07e09b47927c396010442097688ae3c360717f86ab7dc"} Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.525644 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.550877 4677 scope.go:117] "RemoveContainer" containerID="f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.557749 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "22e11417-f484-48b1-a9c6-87b77c7470eb" (UID: "22e11417-f484-48b1-a9c6-87b77c7470eb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.577205 4677 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.577244 4677 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22e11417-f484-48b1-a9c6-87b77c7470eb-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.577255 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.577266 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74cj9\" (UniqueName: \"kubernetes.io/projected/22e11417-f484-48b1-a9c6-87b77c7470eb-kube-api-access-74cj9\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.577280 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.581680 4677 scope.go:117] "RemoveContainer" containerID="9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.581805 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.582101 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11\": container with ID starting with 9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11 not found: ID does not exist" containerID="9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.582132 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11"} err="failed to get container status \"9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11\": rpc error: code = NotFound desc = could not find container \"9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11\": container with ID starting with 9ea6ebe3773f1dbb7c51d2fc462d34d1e7d6995859878666837d1b14c2296d11 not found: ID does not exist" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.582153 4677 scope.go:117] "RemoveContainer" containerID="f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.584499 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "22e11417-f484-48b1-a9c6-87b77c7470eb" (UID: "22e11417-f484-48b1-a9c6-87b77c7470eb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.585466 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab\": container with ID starting with f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab not found: ID does not exist" containerID="f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.585501 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab"} err="failed to get container status \"f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab\": rpc error: code = NotFound desc = could not find container \"f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab\": container with ID starting with f61638d7fbf20af2a5c04e8bb51d4fe0987351b983b55bf971060b77b5790aab not found: ID does not exist" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.585520 4677 scope.go:117] "RemoveContainer" containerID="aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.597383 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.609489 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.609982 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-api" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.609998 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-api" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610011 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="extract-content" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610020 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="extract-content" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610035 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" containerName="nova-manage" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610044 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" containerName="nova-manage" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610065 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="extract-utilities" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610073 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="extract-utilities" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610087 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-log" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610095 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-log" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610104 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerName="init" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610111 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerName="init" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610126 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerName="dnsmasq-dns" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610134 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerName="dnsmasq-dns" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610147 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-metadata" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610155 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-metadata" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610175 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="registry-server" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610214 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="registry-server" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.610233 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-log" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.610244 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-log" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.614245 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-log" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.614293 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ced0c0f-42d8-4695-8f16-310ef7ab0046" containerName="registry-server" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.614314 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" containerName="nova-manage" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.614354 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e74337e-3593-494a-8a23-1b787d670087" containerName="nova-metadata-metadata" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.614375 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-api" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.614400 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d1b2bf-1176-4eb3-8654-f2eb92ad6ddf" containerName="dnsmasq-dns" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.614410 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" containerName="nova-api-log" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.616209 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.618964 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.621493 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.621335 4677 scope.go:117] "RemoveContainer" containerID="95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.622015 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.653284 4677 scope.go:117] "RemoveContainer" containerID="aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.653772 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6\": container with ID starting with aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6 not found: ID does not exist" containerID="aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.653822 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6"} err="failed to get container status \"aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6\": rpc error: code = NotFound desc = could not find container \"aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6\": container with ID starting with aa9cba9176acb5d6f6365b753827be91388e0b4ffd84bc231594eccd4916cdb6 not found: ID does not exist" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.653867 4677 scope.go:117] "RemoveContainer" containerID="95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e" Dec 03 14:13:36 crc kubenswrapper[4677]: E1203 14:13:36.654270 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e\": container with ID starting with 95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e not found: ID does not exist" containerID="95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.654298 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e"} err="failed to get container status \"95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e\": rpc error: code = NotFound desc = could not find container \"95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e\": container with ID starting with 95fef82224c27f81ac456100172a7a8d7fc5f839f57e63673ae34ef39af6401e not found: ID does not exist" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.679153 4677 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22e11417-f484-48b1-a9c6-87b77c7470eb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.780760 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.780872 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-config-data\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.780977 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.781040 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvt6w\" (UniqueName: \"kubernetes.io/projected/9f548e7c-f339-41ce-8cb5-128b48ba2a16-kube-api-access-cvt6w\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.781190 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f548e7c-f339-41ce-8cb5-128b48ba2a16-logs\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.853388 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.863653 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.877017 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.879076 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.880832 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.881409 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.882612 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.882681 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvt6w\" (UniqueName: \"kubernetes.io/projected/9f548e7c-f339-41ce-8cb5-128b48ba2a16-kube-api-access-cvt6w\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.882724 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f548e7c-f339-41ce-8cb5-128b48ba2a16-logs\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.882792 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.882857 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-config-data\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.885057 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f548e7c-f339-41ce-8cb5-128b48ba2a16-logs\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.885163 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.888767 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.890717 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.894692 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.896605 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f548e7c-f339-41ce-8cb5-128b48ba2a16-config-data\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.909707 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvt6w\" (UniqueName: \"kubernetes.io/projected/9f548e7c-f339-41ce-8cb5-128b48ba2a16-kube-api-access-cvt6w\") pod \"nova-metadata-0\" (UID: \"9f548e7c-f339-41ce-8cb5-128b48ba2a16\") " pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.945298 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.985107 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.985222 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-config-data\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.985285 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-public-tls-certs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.985377 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a36537-a4bc-4505-9025-8dfc670f6f0b-logs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.985434 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:36 crc kubenswrapper[4677]: I1203 14:13:36.986035 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpgs6\" (UniqueName: \"kubernetes.io/projected/19a36537-a4bc-4505-9025-8dfc670f6f0b-kube-api-access-tpgs6\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.088235 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-config-data\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.088627 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-public-tls-certs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.088739 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a36537-a4bc-4505-9025-8dfc670f6f0b-logs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.088819 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.088844 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpgs6\" (UniqueName: \"kubernetes.io/projected/19a36537-a4bc-4505-9025-8dfc670f6f0b-kube-api-access-tpgs6\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.088940 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.090074 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a36537-a4bc-4505-9025-8dfc670f6f0b-logs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.093927 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.111386 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-config-data\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.111919 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-public-tls-certs\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.116579 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpgs6\" (UniqueName: \"kubernetes.io/projected/19a36537-a4bc-4505-9025-8dfc670f6f0b-kube-api-access-tpgs6\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.139102 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a36537-a4bc-4505-9025-8dfc670f6f0b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19a36537-a4bc-4505-9025-8dfc670f6f0b\") " pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.351880 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.426267 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.538151 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f548e7c-f339-41ce-8cb5-128b48ba2a16","Type":"ContainerStarted","Data":"616ce28e1098b35a693466ed6c7ed3af6dde118005049f681325ffe54f8646f6"} Dec 03 14:13:37 crc kubenswrapper[4677]: W1203 14:13:37.857344 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19a36537_a4bc_4505_9025_8dfc670f6f0b.slice/crio-798d18a7a99fb06a4664124c0dc4afe696f89eb943276bed4da5965362d2b382 WatchSource:0}: Error finding container 798d18a7a99fb06a4664124c0dc4afe696f89eb943276bed4da5965362d2b382: Status 404 returned error can't find the container with id 798d18a7a99fb06a4664124c0dc4afe696f89eb943276bed4da5965362d2b382 Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.858984 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.996822 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22e11417-f484-48b1-a9c6-87b77c7470eb" path="/var/lib/kubelet/pods/22e11417-f484-48b1-a9c6-87b77c7470eb/volumes" Dec 03 14:13:37 crc kubenswrapper[4677]: I1203 14:13:37.998572 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e74337e-3593-494a-8a23-1b787d670087" path="/var/lib/kubelet/pods/7e74337e-3593-494a-8a23-1b787d670087/volumes" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.437324 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.437918 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.438007 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.439377 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.439465 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" gracePeriod=600 Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.550540 4677 generic.go:334] "Generic (PLEG): container finished" podID="8d042f25-044e-431e-b353-9f37399f41f3" containerID="9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de" exitCode=0 Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.550570 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.550613 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d042f25-044e-431e-b353-9f37399f41f3","Type":"ContainerDied","Data":"9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de"} Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.550645 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d042f25-044e-431e-b353-9f37399f41f3","Type":"ContainerDied","Data":"8a804006adf53aa4a575a28f72d2c84d43b05a9b8f3a46133ca24dfb094cbbbf"} Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.550662 4677 scope.go:117] "RemoveContainer" containerID="9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.559837 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f548e7c-f339-41ce-8cb5-128b48ba2a16","Type":"ContainerStarted","Data":"696a95c55be2d11af4ec64fde6c74562713799bcc31a1734fc5dd3a7e2a67388"} Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.559917 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f548e7c-f339-41ce-8cb5-128b48ba2a16","Type":"ContainerStarted","Data":"25d96228ae9e23a63e9fa12ce4156c8731c97e9ff1ed238fb61256b2710a79d9"} Dec 03 14:13:38 crc kubenswrapper[4677]: E1203 14:13:38.565615 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.573924 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a36537-a4bc-4505-9025-8dfc670f6f0b","Type":"ContainerStarted","Data":"ffd0a0bb02e69b3b7b66c2db05115d4a2388089d37cf27ce444cddbc5fd48da2"} Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.573991 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a36537-a4bc-4505-9025-8dfc670f6f0b","Type":"ContainerStarted","Data":"dddf9faa5c0fa3c7b1a7efa886ffe0726b3ec946697e09fbe2b9ba9cd7e0d8ad"} Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.574003 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a36537-a4bc-4505-9025-8dfc670f6f0b","Type":"ContainerStarted","Data":"798d18a7a99fb06a4664124c0dc4afe696f89eb943276bed4da5965362d2b382"} Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.592825 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.592805257 podStartE2EDuration="2.592805257s" podCreationTimestamp="2025-12-03 14:13:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:13:38.592356666 +0000 UTC m=+1609.338689121" watchObservedRunningTime="2025-12-03 14:13:38.592805257 +0000 UTC m=+1609.339137722" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.605617 4677 scope.go:117] "RemoveContainer" containerID="9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de" Dec 03 14:13:38 crc kubenswrapper[4677]: E1203 14:13:38.606090 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de\": container with ID starting with 9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de not found: ID does not exist" containerID="9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.606118 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de"} err="failed to get container status \"9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de\": rpc error: code = NotFound desc = could not find container \"9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de\": container with ID starting with 9d2a6e61517c844e152ef515052ba67a1b687031a21835a79eeff528025256de not found: ID does not exist" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.630228 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.630208687 podStartE2EDuration="2.630208687s" podCreationTimestamp="2025-12-03 14:13:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:13:38.622704577 +0000 UTC m=+1609.369037052" watchObservedRunningTime="2025-12-03 14:13:38.630208687 +0000 UTC m=+1609.376541142" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.642603 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cktvk\" (UniqueName: \"kubernetes.io/projected/8d042f25-044e-431e-b353-9f37399f41f3-kube-api-access-cktvk\") pod \"8d042f25-044e-431e-b353-9f37399f41f3\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.642713 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-combined-ca-bundle\") pod \"8d042f25-044e-431e-b353-9f37399f41f3\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.642864 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-config-data\") pod \"8d042f25-044e-431e-b353-9f37399f41f3\" (UID: \"8d042f25-044e-431e-b353-9f37399f41f3\") " Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.649457 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d042f25-044e-431e-b353-9f37399f41f3-kube-api-access-cktvk" (OuterVolumeSpecName: "kube-api-access-cktvk") pod "8d042f25-044e-431e-b353-9f37399f41f3" (UID: "8d042f25-044e-431e-b353-9f37399f41f3"). InnerVolumeSpecName "kube-api-access-cktvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.672557 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-config-data" (OuterVolumeSpecName: "config-data") pod "8d042f25-044e-431e-b353-9f37399f41f3" (UID: "8d042f25-044e-431e-b353-9f37399f41f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.687882 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d042f25-044e-431e-b353-9f37399f41f3" (UID: "8d042f25-044e-431e-b353-9f37399f41f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.746018 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cktvk\" (UniqueName: \"kubernetes.io/projected/8d042f25-044e-431e-b353-9f37399f41f3-kube-api-access-cktvk\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.746062 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:38 crc kubenswrapper[4677]: I1203 14:13:38.746073 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d042f25-044e-431e-b353-9f37399f41f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.584346 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.589540 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" exitCode=0 Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.589567 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3"} Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.589609 4677 scope.go:117] "RemoveContainer" containerID="960b2bbd5800fa0b1e3c63b4bd748d2c40afb2ceea7aa3529320dbcc66c17398" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.590694 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:13:39 crc kubenswrapper[4677]: E1203 14:13:39.590979 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.640240 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.655510 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.667006 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:13:39 crc kubenswrapper[4677]: E1203 14:13:39.667517 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d042f25-044e-431e-b353-9f37399f41f3" containerName="nova-scheduler-scheduler" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.667541 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d042f25-044e-431e-b353-9f37399f41f3" containerName="nova-scheduler-scheduler" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.667841 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d042f25-044e-431e-b353-9f37399f41f3" containerName="nova-scheduler-scheduler" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.668752 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.671835 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.679821 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.766865 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a60f05f1-4981-4841-b002-08a655ad4c66-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.766925 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a60f05f1-4981-4841-b002-08a655ad4c66-config-data\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.767074 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdlb5\" (UniqueName: \"kubernetes.io/projected/a60f05f1-4981-4841-b002-08a655ad4c66-kube-api-access-kdlb5\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.868734 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdlb5\" (UniqueName: \"kubernetes.io/projected/a60f05f1-4981-4841-b002-08a655ad4c66-kube-api-access-kdlb5\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.869226 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a60f05f1-4981-4841-b002-08a655ad4c66-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.869346 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a60f05f1-4981-4841-b002-08a655ad4c66-config-data\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.873393 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a60f05f1-4981-4841-b002-08a655ad4c66-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.873524 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a60f05f1-4981-4841-b002-08a655ad4c66-config-data\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.883489 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdlb5\" (UniqueName: \"kubernetes.io/projected/a60f05f1-4981-4841-b002-08a655ad4c66-kube-api-access-kdlb5\") pod \"nova-scheduler-0\" (UID: \"a60f05f1-4981-4841-b002-08a655ad4c66\") " pod="openstack/nova-scheduler-0" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.987628 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d042f25-044e-431e-b353-9f37399f41f3" path="/var/lib/kubelet/pods/8d042f25-044e-431e-b353-9f37399f41f3/volumes" Dec 03 14:13:39 crc kubenswrapper[4677]: I1203 14:13:39.997918 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:13:40 crc kubenswrapper[4677]: I1203 14:13:40.431745 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:13:40 crc kubenswrapper[4677]: I1203 14:13:40.607189 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a60f05f1-4981-4841-b002-08a655ad4c66","Type":"ContainerStarted","Data":"c92dc4a8abc9b074486d5ae163ff4b00f9cd6aa0ee6d3412bd177b1f473d5ceb"} Dec 03 14:13:41 crc kubenswrapper[4677]: I1203 14:13:41.623381 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a60f05f1-4981-4841-b002-08a655ad4c66","Type":"ContainerStarted","Data":"3070117afd554ca2ce160266d71046fc486238c366a77655389c974b03b8ae07"} Dec 03 14:13:41 crc kubenswrapper[4677]: I1203 14:13:41.644894 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.6448679630000003 podStartE2EDuration="2.644867963s" podCreationTimestamp="2025-12-03 14:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:13:41.644351098 +0000 UTC m=+1612.390683603" watchObservedRunningTime="2025-12-03 14:13:41.644867963 +0000 UTC m=+1612.391200438" Dec 03 14:13:41 crc kubenswrapper[4677]: I1203 14:13:41.946072 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:13:41 crc kubenswrapper[4677]: I1203 14:13:41.947639 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:13:45 crc kubenswrapper[4677]: I1203 14:13:44.999750 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 14:13:46 crc kubenswrapper[4677]: I1203 14:13:46.946068 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:13:46 crc kubenswrapper[4677]: I1203 14:13:46.946575 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:13:47 crc kubenswrapper[4677]: I1203 14:13:47.352358 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:13:47 crc kubenswrapper[4677]: I1203 14:13:47.352393 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:13:47 crc kubenswrapper[4677]: I1203 14:13:47.954235 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9f548e7c-f339-41ce-8cb5-128b48ba2a16" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:47 crc kubenswrapper[4677]: I1203 14:13:47.962288 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9f548e7c-f339-41ce-8cb5-128b48ba2a16" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:48 crc kubenswrapper[4677]: I1203 14:13:48.367161 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="19a36537-a4bc-4505-9025-8dfc670f6f0b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.222:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:48 crc kubenswrapper[4677]: I1203 14:13:48.367485 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="19a36537-a4bc-4505-9025-8dfc670f6f0b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.222:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:13:50 crc kubenswrapper[4677]: I1203 14:13:50.006044 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 14:13:50 crc kubenswrapper[4677]: I1203 14:13:50.038615 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 14:13:50 crc kubenswrapper[4677]: I1203 14:13:50.849108 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 14:13:52 crc kubenswrapper[4677]: I1203 14:13:52.977390 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:13:52 crc kubenswrapper[4677]: E1203 14:13:52.978269 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:13:54 crc kubenswrapper[4677]: I1203 14:13:54.735600 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 14:13:56 crc kubenswrapper[4677]: I1203 14:13:56.951712 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:13:56 crc kubenswrapper[4677]: I1203 14:13:56.952596 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:13:56 crc kubenswrapper[4677]: I1203 14:13:56.960273 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:13:57 crc kubenswrapper[4677]: I1203 14:13:57.359266 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:13:57 crc kubenswrapper[4677]: I1203 14:13:57.360085 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:13:57 crc kubenswrapper[4677]: I1203 14:13:57.375875 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:13:57 crc kubenswrapper[4677]: I1203 14:13:57.376008 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:13:57 crc kubenswrapper[4677]: I1203 14:13:57.872659 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:13:57 crc kubenswrapper[4677]: I1203 14:13:57.885338 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:13:57 crc kubenswrapper[4677]: I1203 14:13:57.887630 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:13:58 crc kubenswrapper[4677]: I1203 14:13:58.588676 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:13:58 crc kubenswrapper[4677]: I1203 14:13:58.885268 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" containerName="kube-state-metrics" containerID="cri-o://34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c" gracePeriod=30 Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.377973 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.514789 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2pj9\" (UniqueName: \"kubernetes.io/projected/8afc071f-9f1a-4830-bd22-a1c291e5ae53-kube-api-access-j2pj9\") pod \"8afc071f-9f1a-4830-bd22-a1c291e5ae53\" (UID: \"8afc071f-9f1a-4830-bd22-a1c291e5ae53\") " Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.527755 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8afc071f-9f1a-4830-bd22-a1c291e5ae53-kube-api-access-j2pj9" (OuterVolumeSpecName: "kube-api-access-j2pj9") pod "8afc071f-9f1a-4830-bd22-a1c291e5ae53" (UID: "8afc071f-9f1a-4830-bd22-a1c291e5ae53"). InnerVolumeSpecName "kube-api-access-j2pj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.617539 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2pj9\" (UniqueName: \"kubernetes.io/projected/8afc071f-9f1a-4830-bd22-a1c291e5ae53-kube-api-access-j2pj9\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.897586 4677 generic.go:334] "Generic (PLEG): container finished" podID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" containerID="34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c" exitCode=2 Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.897643 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.897661 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8afc071f-9f1a-4830-bd22-a1c291e5ae53","Type":"ContainerDied","Data":"34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c"} Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.898360 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8afc071f-9f1a-4830-bd22-a1c291e5ae53","Type":"ContainerDied","Data":"d3a01e25e672dea07b5760eae472232cf2e416826d46570dcb7a84499635e560"} Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.898384 4677 scope.go:117] "RemoveContainer" containerID="34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.922056 4677 scope.go:117] "RemoveContainer" containerID="34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c" Dec 03 14:13:59 crc kubenswrapper[4677]: E1203 14:13:59.922592 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c\": container with ID starting with 34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c not found: ID does not exist" containerID="34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.922673 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c"} err="failed to get container status \"34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c\": rpc error: code = NotFound desc = could not find container \"34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c\": container with ID starting with 34fb8c1a6a2f754b8d17fd44e9f4e16f6f1c075293f6a4d23554802410fc036c not found: ID does not exist" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.935295 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.945798 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.956618 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:13:59 crc kubenswrapper[4677]: E1203 14:13:59.957094 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" containerName="kube-state-metrics" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.957110 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" containerName="kube-state-metrics" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.957555 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" containerName="kube-state-metrics" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.958273 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.964477 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.964563 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 14:13:59 crc kubenswrapper[4677]: I1203 14:13:59.970694 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.019711 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8afc071f-9f1a-4830-bd22-a1c291e5ae53" path="/var/lib/kubelet/pods/8afc071f-9f1a-4830-bd22-a1c291e5ae53/volumes" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.129695 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.129751 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.129875 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxb9n\" (UniqueName: \"kubernetes.io/projected/62c96a92-e8b5-4511-8974-a40a064a8267-kube-api-access-xxb9n\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.130080 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.232490 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.232659 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.232682 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.232705 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxb9n\" (UniqueName: \"kubernetes.io/projected/62c96a92-e8b5-4511-8974-a40a064a8267-kube-api-access-xxb9n\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.238985 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.239830 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.240312 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c96a92-e8b5-4511-8974-a40a064a8267-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.250823 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxb9n\" (UniqueName: \"kubernetes.io/projected/62c96a92-e8b5-4511-8974-a40a064a8267-kube-api-access-xxb9n\") pod \"kube-state-metrics-0\" (UID: \"62c96a92-e8b5-4511-8974-a40a064a8267\") " pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.280606 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:14:00 crc kubenswrapper[4677]: W1203 14:14:00.789577 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62c96a92_e8b5_4511_8974_a40a064a8267.slice/crio-b3d3187f620d9cbb22cade9ad7918fd3223e495ed45e991e9a430bc1307785cc WatchSource:0}: Error finding container b3d3187f620d9cbb22cade9ad7918fd3223e495ed45e991e9a430bc1307785cc: Status 404 returned error can't find the container with id b3d3187f620d9cbb22cade9ad7918fd3223e495ed45e991e9a430bc1307785cc Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.794904 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.866090 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.866523 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-central-agent" containerID="cri-o://effea24d757bba312a7308f5cae2de09fead0cec9d7e44bf0d5e9576bd05c038" gracePeriod=30 Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.866600 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="proxy-httpd" containerID="cri-o://9e6b5b8a16aa085236cb2a602bdfc3b1d473ab688a6b259b5fef010f6733f1a1" gracePeriod=30 Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.866600 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="sg-core" containerID="cri-o://73aab7ebcfc7c09953d7083f2578edcbd60ec1347f1a08b6d6970e665f5ddc17" gracePeriod=30 Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.866600 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-notification-agent" containerID="cri-o://7dfe5ffdcda37accb79174c1e77f2afbc04253b47ca875e87e5d58eaa6a6132f" gracePeriod=30 Dec 03 14:14:00 crc kubenswrapper[4677]: I1203 14:14:00.909382 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"62c96a92-e8b5-4511-8974-a40a064a8267","Type":"ContainerStarted","Data":"b3d3187f620d9cbb22cade9ad7918fd3223e495ed45e991e9a430bc1307785cc"} Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.920415 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"62c96a92-e8b5-4511-8974-a40a064a8267","Type":"ContainerStarted","Data":"a203fceef1bb0de7cbc3975fac2fd2faf445be5d2ac4351b3c80bea9cf19a4d1"} Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.920722 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.924642 4677 generic.go:334] "Generic (PLEG): container finished" podID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerID="9e6b5b8a16aa085236cb2a602bdfc3b1d473ab688a6b259b5fef010f6733f1a1" exitCode=0 Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.924663 4677 generic.go:334] "Generic (PLEG): container finished" podID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerID="73aab7ebcfc7c09953d7083f2578edcbd60ec1347f1a08b6d6970e665f5ddc17" exitCode=2 Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.924671 4677 generic.go:334] "Generic (PLEG): container finished" podID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerID="effea24d757bba312a7308f5cae2de09fead0cec9d7e44bf0d5e9576bd05c038" exitCode=0 Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.924683 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerDied","Data":"9e6b5b8a16aa085236cb2a602bdfc3b1d473ab688a6b259b5fef010f6733f1a1"} Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.924748 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerDied","Data":"73aab7ebcfc7c09953d7083f2578edcbd60ec1347f1a08b6d6970e665f5ddc17"} Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.924770 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerDied","Data":"effea24d757bba312a7308f5cae2de09fead0cec9d7e44bf0d5e9576bd05c038"} Dec 03 14:14:01 crc kubenswrapper[4677]: I1203 14:14:01.951992 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.495792423 podStartE2EDuration="2.951939665s" podCreationTimestamp="2025-12-03 14:13:59 +0000 UTC" firstStartedPulling="2025-12-03 14:14:00.79237624 +0000 UTC m=+1631.538708695" lastFinishedPulling="2025-12-03 14:14:01.248523482 +0000 UTC m=+1631.994855937" observedRunningTime="2025-12-03 14:14:01.936124633 +0000 UTC m=+1632.682457118" watchObservedRunningTime="2025-12-03 14:14:01.951939665 +0000 UTC m=+1632.698272160" Dec 03 14:14:02 crc kubenswrapper[4677]: I1203 14:14:02.945104 4677 generic.go:334] "Generic (PLEG): container finished" podID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerID="7dfe5ffdcda37accb79174c1e77f2afbc04253b47ca875e87e5d58eaa6a6132f" exitCode=0 Dec 03 14:14:02 crc kubenswrapper[4677]: I1203 14:14:02.945194 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerDied","Data":"7dfe5ffdcda37accb79174c1e77f2afbc04253b47ca875e87e5d58eaa6a6132f"} Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.145447 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.294866 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-run-httpd\") pod \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.295036 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-sg-core-conf-yaml\") pod \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.295429 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" (UID: "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.297260 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" (UID: "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.295071 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-log-httpd\") pod \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.312172 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-config-data\") pod \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.312294 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-combined-ca-bundle\") pod \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.312338 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm8ql\" (UniqueName: \"kubernetes.io/projected/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-kube-api-access-mm8ql\") pod \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.312382 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-scripts\") pod \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\" (UID: \"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50\") " Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.314389 4677 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.314410 4677 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.358045 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-scripts" (OuterVolumeSpecName: "scripts") pod "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" (UID: "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.359103 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-kube-api-access-mm8ql" (OuterVolumeSpecName: "kube-api-access-mm8ql") pod "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" (UID: "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50"). InnerVolumeSpecName "kube-api-access-mm8ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.373183 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" (UID: "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.416445 4677 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.416481 4677 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.416497 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm8ql\" (UniqueName: \"kubernetes.io/projected/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-kube-api-access-mm8ql\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.423276 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" (UID: "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.444978 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-config-data" (OuterVolumeSpecName: "config-data") pod "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" (UID: "0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.517753 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.518001 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.957262 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50","Type":"ContainerDied","Data":"f6b6669c0e56322d1df2426aadb317d06ae848871239c3e6d039473c7db15c84"} Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.957659 4677 scope.go:117] "RemoveContainer" containerID="9e6b5b8a16aa085236cb2a602bdfc3b1d473ab688a6b259b5fef010f6733f1a1" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.957341 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.975967 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:14:03 crc kubenswrapper[4677]: E1203 14:14:03.976379 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.980854 4677 scope.go:117] "RemoveContainer" containerID="73aab7ebcfc7c09953d7083f2578edcbd60ec1347f1a08b6d6970e665f5ddc17" Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.991285 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:14:03 crc kubenswrapper[4677]: I1203 14:14:03.999553 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.005902 4677 scope.go:117] "RemoveContainer" containerID="7dfe5ffdcda37accb79174c1e77f2afbc04253b47ca875e87e5d58eaa6a6132f" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.019135 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:14:04 crc kubenswrapper[4677]: E1203 14:14:04.019850 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-central-agent" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.019971 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-central-agent" Dec 03 14:14:04 crc kubenswrapper[4677]: E1203 14:14:04.020075 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-notification-agent" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.020588 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-notification-agent" Dec 03 14:14:04 crc kubenswrapper[4677]: E1203 14:14:04.020681 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="sg-core" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.020754 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="sg-core" Dec 03 14:14:04 crc kubenswrapper[4677]: E1203 14:14:04.020831 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="proxy-httpd" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.020897 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="proxy-httpd" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.021395 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-notification-agent" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.021496 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="proxy-httpd" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.021586 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="sg-core" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.021665 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" containerName="ceilometer-central-agent" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.023969 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.026658 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.027075 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.027324 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.051503 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.066395 4677 scope.go:117] "RemoveContainer" containerID="effea24d757bba312a7308f5cae2de09fead0cec9d7e44bf0d5e9576bd05c038" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.128819 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.129225 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-config-data\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.129320 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-scripts\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.129350 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa84d9ae-8578-40d4-a16a-15de08868770-log-httpd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.129469 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26jdd\" (UniqueName: \"kubernetes.io/projected/aa84d9ae-8578-40d4-a16a-15de08868770-kube-api-access-26jdd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.129505 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.130392 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa84d9ae-8578-40d4-a16a-15de08868770-run-httpd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.130445 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.236735 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.237166 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-config-data\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.238338 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-scripts\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.238376 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa84d9ae-8578-40d4-a16a-15de08868770-log-httpd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.238477 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26jdd\" (UniqueName: \"kubernetes.io/projected/aa84d9ae-8578-40d4-a16a-15de08868770-kube-api-access-26jdd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.238509 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.238732 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa84d9ae-8578-40d4-a16a-15de08868770-run-httpd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.238768 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.239171 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa84d9ae-8578-40d4-a16a-15de08868770-log-httpd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.239301 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa84d9ae-8578-40d4-a16a-15de08868770-run-httpd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.242413 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.244915 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-scripts\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.246177 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-config-data\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.253525 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.254161 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa84d9ae-8578-40d4-a16a-15de08868770-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.268890 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26jdd\" (UniqueName: \"kubernetes.io/projected/aa84d9ae-8578-40d4-a16a-15de08868770-kube-api-access-26jdd\") pod \"ceilometer-0\" (UID: \"aa84d9ae-8578-40d4-a16a-15de08868770\") " pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.350381 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:14:04 crc kubenswrapper[4677]: W1203 14:14:04.870419 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa84d9ae_8578_40d4_a16a_15de08868770.slice/crio-8357328a4def45fcfe10d52053288099af4f5f2e4cceb7580d22563fde237453 WatchSource:0}: Error finding container 8357328a4def45fcfe10d52053288099af4f5f2e4cceb7580d22563fde237453: Status 404 returned error can't find the container with id 8357328a4def45fcfe10d52053288099af4f5f2e4cceb7580d22563fde237453 Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.871552 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:14:04 crc kubenswrapper[4677]: I1203 14:14:04.976523 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerStarted","Data":"8357328a4def45fcfe10d52053288099af4f5f2e4cceb7580d22563fde237453"} Dec 03 14:14:05 crc kubenswrapper[4677]: I1203 14:14:05.993425 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50" path="/var/lib/kubelet/pods/0d9d4386-e2d0-4ef1-9be9-1b8e06de3b50/volumes" Dec 03 14:14:05 crc kubenswrapper[4677]: I1203 14:14:05.994718 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerStarted","Data":"9d7daf84846edba6140614669fdd4c4b49383f6c9bd32cbad16651d71733b669"} Dec 03 14:14:05 crc kubenswrapper[4677]: I1203 14:14:05.994747 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerStarted","Data":"95ba62682528829a689d6d975df64208c2ca8799b7aa66d9f8d77e207beeaaf9"} Dec 03 14:14:07 crc kubenswrapper[4677]: I1203 14:14:07.005555 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerStarted","Data":"238f7c9256ce9f8f3737f1f8e2f49e32f23f3ca0ae92c8c7c2491b8d4e0e83ca"} Dec 03 14:14:08 crc kubenswrapper[4677]: I1203 14:14:08.016311 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerStarted","Data":"c15c0c7d8f7b957a893bc69cb35cd6fe597cb766ea28a8e04d102d0072fd8c83"} Dec 03 14:14:08 crc kubenswrapper[4677]: I1203 14:14:08.016735 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:14:08 crc kubenswrapper[4677]: I1203 14:14:08.039326 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.803133631 podStartE2EDuration="5.039286527s" podCreationTimestamp="2025-12-03 14:14:03 +0000 UTC" firstStartedPulling="2025-12-03 14:14:04.873274476 +0000 UTC m=+1635.619606931" lastFinishedPulling="2025-12-03 14:14:07.109427372 +0000 UTC m=+1637.855759827" observedRunningTime="2025-12-03 14:14:08.035416804 +0000 UTC m=+1638.781749269" watchObservedRunningTime="2025-12-03 14:14:08.039286527 +0000 UTC m=+1638.785618982" Dec 03 14:14:08 crc kubenswrapper[4677]: I1203 14:14:08.236447 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:14:09 crc kubenswrapper[4677]: I1203 14:14:09.327615 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:14:10 crc kubenswrapper[4677]: I1203 14:14:10.288766 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 14:14:11 crc kubenswrapper[4677]: I1203 14:14:11.918727 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerName="rabbitmq" containerID="cri-o://2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01" gracePeriod=604797 Dec 03 14:14:12 crc kubenswrapper[4677]: I1203 14:14:12.822177 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerName="rabbitmq" containerID="cri-o://4ecbb69a97190bfc1b6f0687cc33b90e8d85b0f157f94929cbcc8647bf0808fb" gracePeriod=604797 Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.504695 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553240 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-tls\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553342 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-plugins\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553403 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-plugins-conf\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553519 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-config-data\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553657 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-erlang-cookie\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553756 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea48639b-8147-44bb-aa75-96c2ffc0921f-pod-info\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553795 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea48639b-8147-44bb-aa75-96c2ffc0921f-erlang-cookie-secret\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553831 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-confd\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553868 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxgfq\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-kube-api-access-rxgfq\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.553987 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-server-conf\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.554035 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ea48639b-8147-44bb-aa75-96c2ffc0921f\" (UID: \"ea48639b-8147-44bb-aa75-96c2ffc0921f\") " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.555251 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.555358 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.555666 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.555916 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.555933 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.555959 4677 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.563225 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.565176 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-kube-api-access-rxgfq" (OuterVolumeSpecName: "kube-api-access-rxgfq") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "kube-api-access-rxgfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.571170 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ea48639b-8147-44bb-aa75-96c2ffc0921f-pod-info" (OuterVolumeSpecName: "pod-info") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.575143 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.580120 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea48639b-8147-44bb-aa75-96c2ffc0921f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.599942 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-config-data" (OuterVolumeSpecName: "config-data") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.660207 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.660244 4677 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea48639b-8147-44bb-aa75-96c2ffc0921f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.660253 4677 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea48639b-8147-44bb-aa75-96c2ffc0921f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.660263 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxgfq\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-kube-api-access-rxgfq\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.660299 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.660308 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.728596 4677 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.735552 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-server-conf" (OuterVolumeSpecName: "server-conf") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.762236 4677 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea48639b-8147-44bb-aa75-96c2ffc0921f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.762278 4677 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.783137 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ea48639b-8147-44bb-aa75-96c2ffc0921f" (UID: "ea48639b-8147-44bb-aa75-96c2ffc0921f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:13 crc kubenswrapper[4677]: I1203 14:14:13.865208 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea48639b-8147-44bb-aa75-96c2ffc0921f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.154386 4677 generic.go:334] "Generic (PLEG): container finished" podID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerID="4ecbb69a97190bfc1b6f0687cc33b90e8d85b0f157f94929cbcc8647bf0808fb" exitCode=0 Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.154546 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7d5638b0-3c75-4290-b721-d2b5733a0cd3","Type":"ContainerDied","Data":"4ecbb69a97190bfc1b6f0687cc33b90e8d85b0f157f94929cbcc8647bf0808fb"} Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.156805 4677 generic.go:334] "Generic (PLEG): container finished" podID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerID="2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01" exitCode=0 Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.156825 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ea48639b-8147-44bb-aa75-96c2ffc0921f","Type":"ContainerDied","Data":"2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01"} Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.156840 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ea48639b-8147-44bb-aa75-96c2ffc0921f","Type":"ContainerDied","Data":"64dd34abfa008f590cdb828d5463a92c8d0786d45d11451c549d0878ae5339ee"} Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.156855 4677 scope.go:117] "RemoveContainer" containerID="2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.157015 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.220103 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.228126 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.239161 4677 scope.go:117] "RemoveContainer" containerID="52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.247872 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:14:14 crc kubenswrapper[4677]: E1203 14:14:14.248410 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerName="rabbitmq" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.248434 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerName="rabbitmq" Dec 03 14:14:14 crc kubenswrapper[4677]: E1203 14:14:14.248453 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerName="setup-container" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.248462 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerName="setup-container" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.248728 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" containerName="rabbitmq" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.250426 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.253969 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.254139 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.254263 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.254411 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.254608 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.254721 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.258792 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5dkwn" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.277100 4677 scope.go:117] "RemoveContainer" containerID="2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01" Dec 03 14:14:14 crc kubenswrapper[4677]: E1203 14:14:14.278369 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01\": container with ID starting with 2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01 not found: ID does not exist" containerID="2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.278407 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01"} err="failed to get container status \"2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01\": rpc error: code = NotFound desc = could not find container \"2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01\": container with ID starting with 2024f7fd98b17ccb6aca344b616ea9fd69bd66f7f1f99987a01ae645e1630e01 not found: ID does not exist" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.278434 4677 scope.go:117] "RemoveContainer" containerID="52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038" Dec 03 14:14:14 crc kubenswrapper[4677]: E1203 14:14:14.278896 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038\": container with ID starting with 52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038 not found: ID does not exist" containerID="52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.278924 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038"} err="failed to get container status \"52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038\": rpc error: code = NotFound desc = could not find container \"52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038\": container with ID starting with 52cfb72d68b15926e7fb91d2cf5f6dc197bea839a3ca220a499a41cc62cd8038 not found: ID does not exist" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.288554 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.379871 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.379991 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c86t\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-kube-api-access-5c86t\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380036 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380075 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380097 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380172 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380190 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380234 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380310 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380390 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:14 crc kubenswrapper[4677]: I1203 14:14:14.380422 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482306 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482352 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482412 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482468 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c86t\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-kube-api-access-5c86t\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482510 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482550 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482573 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482653 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482674 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482721 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.482760 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.483692 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.483813 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.483879 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.488730 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.488793 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.489563 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.489565 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.489856 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.490810 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.491100 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.513723 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c86t\" (UniqueName: \"kubernetes.io/projected/52fb034c-06d4-4cb2-be6a-0c49a521b0aa-kube-api-access-5c86t\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.561706 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"52fb034c-06d4-4cb2-be6a-0c49a521b0aa\") " pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.637681 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.641806 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790136 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8jf9\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-kube-api-access-g8jf9\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790213 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-tls\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790253 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-confd\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790364 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-server-conf\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790456 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d5638b0-3c75-4290-b721-d2b5733a0cd3-erlang-cookie-secret\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790496 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790518 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-plugins-conf\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790538 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-plugins\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790576 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-erlang-cookie\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790630 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d5638b0-3c75-4290-b721-d2b5733a0cd3-pod-info\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.790692 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-config-data\") pod \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\" (UID: \"7d5638b0-3c75-4290-b721-d2b5733a0cd3\") " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.791165 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.791350 4677 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.794937 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-kube-api-access-g8jf9" (OuterVolumeSpecName: "kube-api-access-g8jf9") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "kube-api-access-g8jf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.796807 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.799003 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.803675 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.810016 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.812360 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7d5638b0-3c75-4290-b721-d2b5733a0cd3-pod-info" (OuterVolumeSpecName: "pod-info") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.831890 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5638b0-3c75-4290-b721-d2b5733a0cd3-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.866979 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-config-data" (OuterVolumeSpecName: "config-data") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.870787 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-server-conf" (OuterVolumeSpecName: "server-conf") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893318 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8jf9\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-kube-api-access-g8jf9\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893349 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893361 4677 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893373 4677 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d5638b0-3c75-4290-b721-d2b5733a0cd3-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893400 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893412 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893424 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893437 4677 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d5638b0-3c75-4290-b721-d2b5733a0cd3-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.893447 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d5638b0-3c75-4290-b721-d2b5733a0cd3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.919450 4677 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.953564 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7d5638b0-3c75-4290-b721-d2b5733a0cd3" (UID: "7d5638b0-3c75-4290-b721-d2b5733a0cd3"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.995242 4677 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:14.995265 4677 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d5638b0-3c75-4290-b721-d2b5733a0cd3-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.182258 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7d5638b0-3c75-4290-b721-d2b5733a0cd3","Type":"ContainerDied","Data":"1eaca61478414b41173979934f53b8d704256a03bee66f24335b8368519ea10c"} Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.182313 4677 scope.go:117] "RemoveContainer" containerID="4ecbb69a97190bfc1b6f0687cc33b90e8d85b0f157f94929cbcc8647bf0808fb" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.182270 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.245611 4677 scope.go:117] "RemoveContainer" containerID="c6d03f88aa96a8f0d3893995f7b8086caba5f3f4e8991faae35b447b28f9527a" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.265899 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.287938 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.310283 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:14:15 crc kubenswrapper[4677]: E1203 14:14:15.310808 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerName="setup-container" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.310827 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerName="setup-container" Dec 03 14:14:15 crc kubenswrapper[4677]: E1203 14:14:15.310857 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerName="rabbitmq" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.310863 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerName="rabbitmq" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.311064 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" containerName="rabbitmq" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.312322 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.320830 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.321094 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.321099 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.321306 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.321365 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.321457 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.321511 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vqzcn" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.336137 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.402193 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.402507 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.402610 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29df9745-7bd4-428e-9714-624cf6f66d7e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.402887 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29df9745-7bd4-428e-9714-624cf6f66d7e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.402942 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.402988 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ps4g\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-kube-api-access-7ps4g\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.403175 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.403291 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.403381 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.403468 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.403559 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: W1203 14:14:15.403892 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52fb034c_06d4_4cb2_be6a_0c49a521b0aa.slice/crio-a245cc0ff9945254d304f0283c85b694ed8bfb771857c54c689ae00469b27173 WatchSource:0}: Error finding container a245cc0ff9945254d304f0283c85b694ed8bfb771857c54c689ae00469b27173: Status 404 returned error can't find the container with id a245cc0ff9945254d304f0283c85b694ed8bfb771857c54c689ae00469b27173 Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.411174 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.506745 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.507076 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.507137 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.507264 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.507365 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29df9745-7bd4-428e-9714-624cf6f66d7e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.507842 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.508640 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29df9745-7bd4-428e-9714-624cf6f66d7e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.508676 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.509023 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ps4g\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-kube-api-access-7ps4g\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.509245 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.509273 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.509386 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.509524 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.510113 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.510476 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.510698 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.513314 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29df9745-7bd4-428e-9714-624cf6f66d7e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.517643 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29df9745-7bd4-428e-9714-624cf6f66d7e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.519541 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.519846 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.523795 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29df9745-7bd4-428e-9714-624cf6f66d7e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.527202 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ps4g\" (UniqueName: \"kubernetes.io/projected/29df9745-7bd4-428e-9714-624cf6f66d7e-kube-api-access-7ps4g\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.549736 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29df9745-7bd4-428e-9714-624cf6f66d7e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.643457 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.976972 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:14:15 crc kubenswrapper[4677]: E1203 14:14:15.977733 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.992133 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d5638b0-3c75-4290-b721-d2b5733a0cd3" path="/var/lib/kubelet/pods/7d5638b0-3c75-4290-b721-d2b5733a0cd3/volumes" Dec 03 14:14:15 crc kubenswrapper[4677]: I1203 14:14:15.993437 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea48639b-8147-44bb-aa75-96c2ffc0921f" path="/var/lib/kubelet/pods/ea48639b-8147-44bb-aa75-96c2ffc0921f/volumes" Dec 03 14:14:16 crc kubenswrapper[4677]: W1203 14:14:16.072797 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29df9745_7bd4_428e_9714_624cf6f66d7e.slice/crio-429f925d15181ec3001e4c6e6d05f2307a6210a73c18d89888c19035f7df30be WatchSource:0}: Error finding container 429f925d15181ec3001e4c6e6d05f2307a6210a73c18d89888c19035f7df30be: Status 404 returned error can't find the container with id 429f925d15181ec3001e4c6e6d05f2307a6210a73c18d89888c19035f7df30be Dec 03 14:14:16 crc kubenswrapper[4677]: I1203 14:14:16.074069 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:14:16 crc kubenswrapper[4677]: I1203 14:14:16.237840 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52fb034c-06d4-4cb2-be6a-0c49a521b0aa","Type":"ContainerStarted","Data":"a245cc0ff9945254d304f0283c85b694ed8bfb771857c54c689ae00469b27173"} Dec 03 14:14:16 crc kubenswrapper[4677]: I1203 14:14:16.240010 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29df9745-7bd4-428e-9714-624cf6f66d7e","Type":"ContainerStarted","Data":"429f925d15181ec3001e4c6e6d05f2307a6210a73c18d89888c19035f7df30be"} Dec 03 14:14:18 crc kubenswrapper[4677]: I1203 14:14:18.286650 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52fb034c-06d4-4cb2-be6a-0c49a521b0aa","Type":"ContainerStarted","Data":"fb521d1bfed620f67dc6adf4c8a226102b0995a967744b4c1b7204175b5323bf"} Dec 03 14:14:18 crc kubenswrapper[4677]: I1203 14:14:18.298685 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29df9745-7bd4-428e-9714-624cf6f66d7e","Type":"ContainerStarted","Data":"d0485ff8a8fc19bf9d5244d3d7a43650ef11a98f9868f676b7b24d6796644bd3"} Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.864155 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69b769dc59-mdq7q"] Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.866747 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.879723 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.886701 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b769dc59-mdq7q"] Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.977177 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-sb\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.977220 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-swift-storage-0\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.977257 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-nb\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.977334 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-svc\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.977401 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrpwh\" (UniqueName: \"kubernetes.io/projected/077c7d38-93a0-47fa-9bea-bdae662f553a-kube-api-access-qrpwh\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.977436 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-openstack-edpm-ipam\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:23 crc kubenswrapper[4677]: I1203 14:14:23.977476 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-config\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.079524 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-openstack-edpm-ipam\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.079595 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-config\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.079686 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-sb\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.079709 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-swift-storage-0\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.079744 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-nb\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.079785 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-svc\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.079897 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrpwh\" (UniqueName: \"kubernetes.io/projected/077c7d38-93a0-47fa-9bea-bdae662f553a-kube-api-access-qrpwh\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.081015 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-nb\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.081047 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-config\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.081096 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-swift-storage-0\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.081125 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-svc\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.081450 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-openstack-edpm-ipam\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.081509 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-sb\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.103709 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrpwh\" (UniqueName: \"kubernetes.io/projected/077c7d38-93a0-47fa-9bea-bdae662f553a-kube-api-access-qrpwh\") pod \"dnsmasq-dns-69b769dc59-mdq7q\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.190046 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:24 crc kubenswrapper[4677]: I1203 14:14:24.645752 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b769dc59-mdq7q"] Dec 03 14:14:24 crc kubenswrapper[4677]: W1203 14:14:24.646532 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod077c7d38_93a0_47fa_9bea_bdae662f553a.slice/crio-f002b8a2fc10a9190cae2fdb650dd3afb969b282bb14ae9375b54fc4fcddfef2 WatchSource:0}: Error finding container f002b8a2fc10a9190cae2fdb650dd3afb969b282bb14ae9375b54fc4fcddfef2: Status 404 returned error can't find the container with id f002b8a2fc10a9190cae2fdb650dd3afb969b282bb14ae9375b54fc4fcddfef2 Dec 03 14:14:25 crc kubenswrapper[4677]: I1203 14:14:25.368013 4677 generic.go:334] "Generic (PLEG): container finished" podID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerID="9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d" exitCode=0 Dec 03 14:14:25 crc kubenswrapper[4677]: I1203 14:14:25.368087 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" event={"ID":"077c7d38-93a0-47fa-9bea-bdae662f553a","Type":"ContainerDied","Data":"9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d"} Dec 03 14:14:25 crc kubenswrapper[4677]: I1203 14:14:25.368602 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" event={"ID":"077c7d38-93a0-47fa-9bea-bdae662f553a","Type":"ContainerStarted","Data":"f002b8a2fc10a9190cae2fdb650dd3afb969b282bb14ae9375b54fc4fcddfef2"} Dec 03 14:14:26 crc kubenswrapper[4677]: I1203 14:14:26.383993 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" event={"ID":"077c7d38-93a0-47fa-9bea-bdae662f553a","Type":"ContainerStarted","Data":"55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db"} Dec 03 14:14:26 crc kubenswrapper[4677]: I1203 14:14:26.384555 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:26 crc kubenswrapper[4677]: I1203 14:14:26.408024 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" podStartSLOduration=3.408001684 podStartE2EDuration="3.408001684s" podCreationTimestamp="2025-12-03 14:14:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:26.401212883 +0000 UTC m=+1657.147545338" watchObservedRunningTime="2025-12-03 14:14:26.408001684 +0000 UTC m=+1657.154334139" Dec 03 14:14:29 crc kubenswrapper[4677]: I1203 14:14:29.989720 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:14:29 crc kubenswrapper[4677]: E1203 14:14:29.990527 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.192228 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.279094 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-846d784fbc-wmfdq"] Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.279363 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" podUID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerName="dnsmasq-dns" containerID="cri-o://02f7b8f50c4afc59b154aee5813eecbde9a5e88ee91aa969c37d05cf52d1c40e" gracePeriod=10 Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.374789 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.464704 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-574fbf444c-q9lfl"] Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.466866 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.481685 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574fbf444c-q9lfl"] Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.487375 4677 generic.go:334] "Generic (PLEG): container finished" podID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerID="02f7b8f50c4afc59b154aee5813eecbde9a5e88ee91aa969c37d05cf52d1c40e" exitCode=0 Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.487412 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" event={"ID":"7b69f0ff-660e-4535-8d71-10b8d67dfb2b","Type":"ContainerDied","Data":"02f7b8f50c4afc59b154aee5813eecbde9a5e88ee91aa969c37d05cf52d1c40e"} Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.620362 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-dns-svc\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.620418 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzczs\" (UniqueName: \"kubernetes.io/projected/60649fbd-8fcf-48d2-9327-6aef75ca6163-kube-api-access-rzczs\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.620461 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-openstack-edpm-ipam\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.620793 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-ovsdbserver-sb\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.620903 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-ovsdbserver-nb\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.620968 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-dns-swift-storage-0\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.621033 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-config\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.723125 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-config\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.723241 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-dns-svc\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.723275 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzczs\" (UniqueName: \"kubernetes.io/projected/60649fbd-8fcf-48d2-9327-6aef75ca6163-kube-api-access-rzczs\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.723328 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-openstack-edpm-ipam\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.723391 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-ovsdbserver-sb\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.723420 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-ovsdbserver-nb\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.723449 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-dns-swift-storage-0\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.724565 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-dns-swift-storage-0\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.724631 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-ovsdbserver-sb\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.725047 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-ovsdbserver-nb\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.725339 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-dns-svc\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.725373 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-openstack-edpm-ipam\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.726533 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60649fbd-8fcf-48d2-9327-6aef75ca6163-config\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.746988 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzczs\" (UniqueName: \"kubernetes.io/projected/60649fbd-8fcf-48d2-9327-6aef75ca6163-kube-api-access-rzczs\") pod \"dnsmasq-dns-574fbf444c-q9lfl\" (UID: \"60649fbd-8fcf-48d2-9327-6aef75ca6163\") " pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.796106 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:34 crc kubenswrapper[4677]: I1203 14:14:34.935246 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.028552 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-swift-storage-0\") pod \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.028638 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-svc\") pod \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.028669 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-config\") pod \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.028720 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-nb\") pod \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.028794 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-sb\") pod \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.028936 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46ntn\" (UniqueName: \"kubernetes.io/projected/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-kube-api-access-46ntn\") pod \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\" (UID: \"7b69f0ff-660e-4535-8d71-10b8d67dfb2b\") " Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.058294 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-kube-api-access-46ntn" (OuterVolumeSpecName: "kube-api-access-46ntn") pod "7b69f0ff-660e-4535-8d71-10b8d67dfb2b" (UID: "7b69f0ff-660e-4535-8d71-10b8d67dfb2b"). InnerVolumeSpecName "kube-api-access-46ntn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.126625 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7b69f0ff-660e-4535-8d71-10b8d67dfb2b" (UID: "7b69f0ff-660e-4535-8d71-10b8d67dfb2b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.131118 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46ntn\" (UniqueName: \"kubernetes.io/projected/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-kube-api-access-46ntn\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.131145 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.143481 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7b69f0ff-660e-4535-8d71-10b8d67dfb2b" (UID: "7b69f0ff-660e-4535-8d71-10b8d67dfb2b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.160730 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7b69f0ff-660e-4535-8d71-10b8d67dfb2b" (UID: "7b69f0ff-660e-4535-8d71-10b8d67dfb2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.163420 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7b69f0ff-660e-4535-8d71-10b8d67dfb2b" (UID: "7b69f0ff-660e-4535-8d71-10b8d67dfb2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.183579 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-config" (OuterVolumeSpecName: "config") pod "7b69f0ff-660e-4535-8d71-10b8d67dfb2b" (UID: "7b69f0ff-660e-4535-8d71-10b8d67dfb2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.236971 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.237007 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.237019 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.237029 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b69f0ff-660e-4535-8d71-10b8d67dfb2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.441816 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574fbf444c-q9lfl"] Dec 03 14:14:35 crc kubenswrapper[4677]: W1203 14:14:35.446247 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60649fbd_8fcf_48d2_9327_6aef75ca6163.slice/crio-3332513d0bcce4bc737855d8399e9dfe085377c15ef7b32ca4330b8a7eb32289 WatchSource:0}: Error finding container 3332513d0bcce4bc737855d8399e9dfe085377c15ef7b32ca4330b8a7eb32289: Status 404 returned error can't find the container with id 3332513d0bcce4bc737855d8399e9dfe085377c15ef7b32ca4330b8a7eb32289 Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.525248 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" event={"ID":"60649fbd-8fcf-48d2-9327-6aef75ca6163","Type":"ContainerStarted","Data":"3332513d0bcce4bc737855d8399e9dfe085377c15ef7b32ca4330b8a7eb32289"} Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.527098 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" event={"ID":"7b69f0ff-660e-4535-8d71-10b8d67dfb2b","Type":"ContainerDied","Data":"decaee75b304bc99129e0c7d31073862879e946d0bc3556d3c6445ec8df5d3a3"} Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.527130 4677 scope.go:117] "RemoveContainer" containerID="02f7b8f50c4afc59b154aee5813eecbde9a5e88ee91aa969c37d05cf52d1c40e" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.527318 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-846d784fbc-wmfdq" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.611159 4677 scope.go:117] "RemoveContainer" containerID="2a95e17880c372c0bc5379a81c445f983175075784281af1c5de4f1f31901ad6" Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.657476 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-846d784fbc-wmfdq"] Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.666432 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-846d784fbc-wmfdq"] Dec 03 14:14:35 crc kubenswrapper[4677]: I1203 14:14:35.990194 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" path="/var/lib/kubelet/pods/7b69f0ff-660e-4535-8d71-10b8d67dfb2b/volumes" Dec 03 14:14:36 crc kubenswrapper[4677]: I1203 14:14:36.538933 4677 generic.go:334] "Generic (PLEG): container finished" podID="60649fbd-8fcf-48d2-9327-6aef75ca6163" containerID="079c337325c697cb5d20c0a38cf79d053125b29d1a3b66d9f831e4c71007616c" exitCode=0 Dec 03 14:14:36 crc kubenswrapper[4677]: I1203 14:14:36.539029 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" event={"ID":"60649fbd-8fcf-48d2-9327-6aef75ca6163","Type":"ContainerDied","Data":"079c337325c697cb5d20c0a38cf79d053125b29d1a3b66d9f831e4c71007616c"} Dec 03 14:14:37 crc kubenswrapper[4677]: I1203 14:14:37.558586 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" event={"ID":"60649fbd-8fcf-48d2-9327-6aef75ca6163","Type":"ContainerStarted","Data":"f71ea3a83c810d9e29ad7eb73ac9e4f269ce820a10316be3817e5994242ecedd"} Dec 03 14:14:37 crc kubenswrapper[4677]: I1203 14:14:37.559416 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:37 crc kubenswrapper[4677]: I1203 14:14:37.582270 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" podStartSLOduration=3.582248658 podStartE2EDuration="3.582248658s" podCreationTimestamp="2025-12-03 14:14:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:37.5774539 +0000 UTC m=+1668.323786385" watchObservedRunningTime="2025-12-03 14:14:37.582248658 +0000 UTC m=+1668.328581113" Dec 03 14:14:40 crc kubenswrapper[4677]: I1203 14:14:40.975724 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:14:40 crc kubenswrapper[4677]: E1203 14:14:40.976420 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:14:44 crc kubenswrapper[4677]: I1203 14:14:44.798168 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-574fbf444c-q9lfl" Dec 03 14:14:44 crc kubenswrapper[4677]: I1203 14:14:44.864068 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69b769dc59-mdq7q"] Dec 03 14:14:44 crc kubenswrapper[4677]: I1203 14:14:44.864359 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" podUID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerName="dnsmasq-dns" containerID="cri-o://55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db" gracePeriod=10 Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.383289 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.453901 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-swift-storage-0\") pod \"077c7d38-93a0-47fa-9bea-bdae662f553a\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.453976 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrpwh\" (UniqueName: \"kubernetes.io/projected/077c7d38-93a0-47fa-9bea-bdae662f553a-kube-api-access-qrpwh\") pod \"077c7d38-93a0-47fa-9bea-bdae662f553a\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.454087 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-sb\") pod \"077c7d38-93a0-47fa-9bea-bdae662f553a\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.454132 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-config\") pod \"077c7d38-93a0-47fa-9bea-bdae662f553a\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.454228 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-svc\") pod \"077c7d38-93a0-47fa-9bea-bdae662f553a\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.454298 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-nb\") pod \"077c7d38-93a0-47fa-9bea-bdae662f553a\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.454359 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-openstack-edpm-ipam\") pod \"077c7d38-93a0-47fa-9bea-bdae662f553a\" (UID: \"077c7d38-93a0-47fa-9bea-bdae662f553a\") " Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.470301 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/077c7d38-93a0-47fa-9bea-bdae662f553a-kube-api-access-qrpwh" (OuterVolumeSpecName: "kube-api-access-qrpwh") pod "077c7d38-93a0-47fa-9bea-bdae662f553a" (UID: "077c7d38-93a0-47fa-9bea-bdae662f553a"). InnerVolumeSpecName "kube-api-access-qrpwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.506595 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "077c7d38-93a0-47fa-9bea-bdae662f553a" (UID: "077c7d38-93a0-47fa-9bea-bdae662f553a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.510475 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "077c7d38-93a0-47fa-9bea-bdae662f553a" (UID: "077c7d38-93a0-47fa-9bea-bdae662f553a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.511772 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "077c7d38-93a0-47fa-9bea-bdae662f553a" (UID: "077c7d38-93a0-47fa-9bea-bdae662f553a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.513167 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "077c7d38-93a0-47fa-9bea-bdae662f553a" (UID: "077c7d38-93a0-47fa-9bea-bdae662f553a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.520184 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-config" (OuterVolumeSpecName: "config") pod "077c7d38-93a0-47fa-9bea-bdae662f553a" (UID: "077c7d38-93a0-47fa-9bea-bdae662f553a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.522464 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "077c7d38-93a0-47fa-9bea-bdae662f553a" (UID: "077c7d38-93a0-47fa-9bea-bdae662f553a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.556701 4677 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.556731 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrpwh\" (UniqueName: \"kubernetes.io/projected/077c7d38-93a0-47fa-9bea-bdae662f553a-kube-api-access-qrpwh\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.556743 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.556755 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.556764 4677 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.556772 4677 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.556780 4677 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/077c7d38-93a0-47fa-9bea-bdae662f553a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.639496 4677 generic.go:334] "Generic (PLEG): container finished" podID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerID="55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db" exitCode=0 Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.639553 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" event={"ID":"077c7d38-93a0-47fa-9bea-bdae662f553a","Type":"ContainerDied","Data":"55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db"} Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.639584 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" event={"ID":"077c7d38-93a0-47fa-9bea-bdae662f553a","Type":"ContainerDied","Data":"f002b8a2fc10a9190cae2fdb650dd3afb969b282bb14ae9375b54fc4fcddfef2"} Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.639604 4677 scope.go:117] "RemoveContainer" containerID="55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.639605 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b769dc59-mdq7q" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.660869 4677 scope.go:117] "RemoveContainer" containerID="9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.688612 4677 scope.go:117] "RemoveContainer" containerID="55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db" Dec 03 14:14:45 crc kubenswrapper[4677]: E1203 14:14:45.689476 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db\": container with ID starting with 55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db not found: ID does not exist" containerID="55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.689673 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db"} err="failed to get container status \"55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db\": rpc error: code = NotFound desc = could not find container \"55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db\": container with ID starting with 55277cbe78d145be4dc65133da7fb7d5397a4a5e854ab735f9a007055d50d6db not found: ID does not exist" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.690130 4677 scope.go:117] "RemoveContainer" containerID="9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d" Dec 03 14:14:45 crc kubenswrapper[4677]: E1203 14:14:45.690687 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d\": container with ID starting with 9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d not found: ID does not exist" containerID="9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.690811 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d"} err="failed to get container status \"9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d\": rpc error: code = NotFound desc = could not find container \"9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d\": container with ID starting with 9819d30693ff6bcd07631ac8129b6eaffde27049719072fa55ddb5ab7415ea1d not found: ID does not exist" Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.700746 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69b769dc59-mdq7q"] Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.714026 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69b769dc59-mdq7q"] Dec 03 14:14:45 crc kubenswrapper[4677]: I1203 14:14:45.989556 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="077c7d38-93a0-47fa-9bea-bdae662f553a" path="/var/lib/kubelet/pods/077c7d38-93a0-47fa-9bea-bdae662f553a/volumes" Dec 03 14:14:50 crc kubenswrapper[4677]: I1203 14:14:50.692285 4677 generic.go:334] "Generic (PLEG): container finished" podID="52fb034c-06d4-4cb2-be6a-0c49a521b0aa" containerID="fb521d1bfed620f67dc6adf4c8a226102b0995a967744b4c1b7204175b5323bf" exitCode=0 Dec 03 14:14:50 crc kubenswrapper[4677]: I1203 14:14:50.692400 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52fb034c-06d4-4cb2-be6a-0c49a521b0aa","Type":"ContainerDied","Data":"fb521d1bfed620f67dc6adf4c8a226102b0995a967744b4c1b7204175b5323bf"} Dec 03 14:14:50 crc kubenswrapper[4677]: I1203 14:14:50.697509 4677 generic.go:334] "Generic (PLEG): container finished" podID="29df9745-7bd4-428e-9714-624cf6f66d7e" containerID="d0485ff8a8fc19bf9d5244d3d7a43650ef11a98f9868f676b7b24d6796644bd3" exitCode=0 Dec 03 14:14:50 crc kubenswrapper[4677]: I1203 14:14:50.697606 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29df9745-7bd4-428e-9714-624cf6f66d7e","Type":"ContainerDied","Data":"d0485ff8a8fc19bf9d5244d3d7a43650ef11a98f9868f676b7b24d6796644bd3"} Dec 03 14:14:51 crc kubenswrapper[4677]: I1203 14:14:51.708932 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29df9745-7bd4-428e-9714-624cf6f66d7e","Type":"ContainerStarted","Data":"11302ca77d0e2469e08b9574a0c451adb29432ea230666b3febabdfd591f7489"} Dec 03 14:14:51 crc kubenswrapper[4677]: I1203 14:14:51.709478 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:14:51 crc kubenswrapper[4677]: I1203 14:14:51.711005 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52fb034c-06d4-4cb2-be6a-0c49a521b0aa","Type":"ContainerStarted","Data":"c7a577f99557933534b542e9d966a4540f629e0f1982a38f876675e23371b084"} Dec 03 14:14:51 crc kubenswrapper[4677]: I1203 14:14:51.712162 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 14:14:51 crc kubenswrapper[4677]: I1203 14:14:51.742114 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.742098107 podStartE2EDuration="36.742098107s" podCreationTimestamp="2025-12-03 14:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:51.734295328 +0000 UTC m=+1682.480627803" watchObservedRunningTime="2025-12-03 14:14:51.742098107 +0000 UTC m=+1682.488430562" Dec 03 14:14:52 crc kubenswrapper[4677]: I1203 14:14:52.975935 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:14:52 crc kubenswrapper[4677]: E1203 14:14:52.976216 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.128160 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=46.1281384 podStartE2EDuration="46.1281384s" podCreationTimestamp="2025-12-03 14:14:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:51.766846799 +0000 UTC m=+1682.513179274" watchObservedRunningTime="2025-12-03 14:15:00.1281384 +0000 UTC m=+1690.874470865" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.139668 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt"] Dec 03 14:15:00 crc kubenswrapper[4677]: E1203 14:15:00.140316 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerName="dnsmasq-dns" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.140384 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerName="dnsmasq-dns" Dec 03 14:15:00 crc kubenswrapper[4677]: E1203 14:15:00.140449 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerName="init" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.140511 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerName="init" Dec 03 14:15:00 crc kubenswrapper[4677]: E1203 14:15:00.140692 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerName="init" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.140782 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerName="init" Dec 03 14:15:00 crc kubenswrapper[4677]: E1203 14:15:00.140861 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerName="dnsmasq-dns" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.140965 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerName="dnsmasq-dns" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.141285 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b69f0ff-660e-4535-8d71-10b8d67dfb2b" containerName="dnsmasq-dns" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.141360 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="077c7d38-93a0-47fa-9bea-bdae662f553a" containerName="dnsmasq-dns" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.142121 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.145992 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.146371 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.158435 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt"] Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.245733 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-secret-volume\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.245800 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcpqm\" (UniqueName: \"kubernetes.io/projected/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-kube-api-access-mcpqm\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.245869 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-config-volume\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.348380 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-secret-volume\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.348441 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcpqm\" (UniqueName: \"kubernetes.io/projected/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-kube-api-access-mcpqm\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.348522 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-config-volume\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.349573 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-config-volume\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.354093 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-secret-volume\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.368074 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcpqm\" (UniqueName: \"kubernetes.io/projected/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-kube-api-access-mcpqm\") pod \"collect-profiles-29412855-s44zt\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.464381 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:00 crc kubenswrapper[4677]: I1203 14:15:00.948496 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt"] Dec 03 14:15:00 crc kubenswrapper[4677]: W1203 14:15:00.953437 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94a4aeaa_b2e5_4d62_9c8d_be9774befcaf.slice/crio-0461f1352dc862d90ea98551243e3845c243b94350406a1b3883ef43819f47f2 WatchSource:0}: Error finding container 0461f1352dc862d90ea98551243e3845c243b94350406a1b3883ef43819f47f2: Status 404 returned error can't find the container with id 0461f1352dc862d90ea98551243e3845c243b94350406a1b3883ef43819f47f2 Dec 03 14:15:01 crc kubenswrapper[4677]: I1203 14:15:01.798699 4677 generic.go:334] "Generic (PLEG): container finished" podID="94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" containerID="aeb7d313a45f0f2de4ec4a6bd4786ab579ca0bee5aa3482accfd3b608a2445b6" exitCode=0 Dec 03 14:15:01 crc kubenswrapper[4677]: I1203 14:15:01.798752 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" event={"ID":"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf","Type":"ContainerDied","Data":"aeb7d313a45f0f2de4ec4a6bd4786ab579ca0bee5aa3482accfd3b608a2445b6"} Dec 03 14:15:01 crc kubenswrapper[4677]: I1203 14:15:01.799080 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" event={"ID":"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf","Type":"ContainerStarted","Data":"0461f1352dc862d90ea98551243e3845c243b94350406a1b3883ef43819f47f2"} Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.050114 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv"] Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.051839 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.054188 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.055187 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.056264 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.056556 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.078706 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv"] Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.108348 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.108480 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.108649 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ktkr\" (UniqueName: \"kubernetes.io/projected/c2dc4bf3-fecf-4b2a-a664-d5c543196286-kube-api-access-5ktkr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.108835 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.183327 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.210320 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.210413 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.210503 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ktkr\" (UniqueName: \"kubernetes.io/projected/c2dc4bf3-fecf-4b2a-a664-d5c543196286-kube-api-access-5ktkr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.210581 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.215820 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.220591 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.232383 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.236804 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ktkr\" (UniqueName: \"kubernetes.io/projected/c2dc4bf3-fecf-4b2a-a664-d5c543196286-kube-api-access-5ktkr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.311431 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcpqm\" (UniqueName: \"kubernetes.io/projected/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-kube-api-access-mcpqm\") pod \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.311545 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-config-volume\") pod \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.311586 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-secret-volume\") pod \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\" (UID: \"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf\") " Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.312438 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-config-volume" (OuterVolumeSpecName: "config-volume") pod "94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" (UID: "94a4aeaa-b2e5-4d62-9c8d-be9774befcaf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.314657 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" (UID: "94a4aeaa-b2e5-4d62-9c8d-be9774befcaf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.315389 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-kube-api-access-mcpqm" (OuterVolumeSpecName: "kube-api-access-mcpqm") pod "94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" (UID: "94a4aeaa-b2e5-4d62-9c8d-be9774befcaf"). InnerVolumeSpecName "kube-api-access-mcpqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.371574 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.414137 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcpqm\" (UniqueName: \"kubernetes.io/projected/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-kube-api-access-mcpqm\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.414182 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.414213 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.817960 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" event={"ID":"94a4aeaa-b2e5-4d62-9c8d-be9774befcaf","Type":"ContainerDied","Data":"0461f1352dc862d90ea98551243e3845c243b94350406a1b3883ef43819f47f2"} Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.818297 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0461f1352dc862d90ea98551243e3845c243b94350406a1b3883ef43819f47f2" Dec 03 14:15:03 crc kubenswrapper[4677]: I1203 14:15:03.818191 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt" Dec 03 14:15:04 crc kubenswrapper[4677]: I1203 14:15:04.008676 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv"] Dec 03 14:15:04 crc kubenswrapper[4677]: W1203 14:15:04.011409 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2dc4bf3_fecf_4b2a_a664_d5c543196286.slice/crio-7adafff7923900473a3c1c9ffe828b3b4c14abec48e7a46e579594c5e849334a WatchSource:0}: Error finding container 7adafff7923900473a3c1c9ffe828b3b4c14abec48e7a46e579594c5e849334a: Status 404 returned error can't find the container with id 7adafff7923900473a3c1c9ffe828b3b4c14abec48e7a46e579594c5e849334a Dec 03 14:15:04 crc kubenswrapper[4677]: I1203 14:15:04.446471 4677 scope.go:117] "RemoveContainer" containerID="c3e5a3fab9ee33457e644fbf29a56ed57ceb2fc629f3a944fa9b2cf28c4d4933" Dec 03 14:15:04 crc kubenswrapper[4677]: I1203 14:15:04.641168 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 14:15:04 crc kubenswrapper[4677]: I1203 14:15:04.843183 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" event={"ID":"c2dc4bf3-fecf-4b2a-a664-d5c543196286","Type":"ContainerStarted","Data":"7adafff7923900473a3c1c9ffe828b3b4c14abec48e7a46e579594c5e849334a"} Dec 03 14:15:04 crc kubenswrapper[4677]: I1203 14:15:04.976292 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:15:04 crc kubenswrapper[4677]: E1203 14:15:04.976750 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:15:05 crc kubenswrapper[4677]: I1203 14:15:05.649151 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:15:13 crc kubenswrapper[4677]: I1203 14:15:13.949446 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" event={"ID":"c2dc4bf3-fecf-4b2a-a664-d5c543196286","Type":"ContainerStarted","Data":"9d309ca1a878a0406de51d6acdf7a0d147f15efbe7ef7d167c8f146d4f3f00ca"} Dec 03 14:15:13 crc kubenswrapper[4677]: I1203 14:15:13.977696 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" podStartSLOduration=1.7814887769999999 podStartE2EDuration="10.977674981s" podCreationTimestamp="2025-12-03 14:15:03 +0000 UTC" firstStartedPulling="2025-12-03 14:15:04.014363234 +0000 UTC m=+1694.760695689" lastFinishedPulling="2025-12-03 14:15:13.210549428 +0000 UTC m=+1703.956881893" observedRunningTime="2025-12-03 14:15:13.97347682 +0000 UTC m=+1704.719809275" watchObservedRunningTime="2025-12-03 14:15:13.977674981 +0000 UTC m=+1704.724007446" Dec 03 14:15:18 crc kubenswrapper[4677]: I1203 14:15:18.212677 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:15:18 crc kubenswrapper[4677]: E1203 14:15:18.215190 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:15:25 crc kubenswrapper[4677]: E1203 14:15:25.402435 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2dc4bf3_fecf_4b2a_a664_d5c543196286.slice/crio-conmon-9d309ca1a878a0406de51d6acdf7a0d147f15efbe7ef7d167c8f146d4f3f00ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2dc4bf3_fecf_4b2a_a664_d5c543196286.slice/crio-9d309ca1a878a0406de51d6acdf7a0d147f15efbe7ef7d167c8f146d4f3f00ca.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:15:26 crc kubenswrapper[4677]: I1203 14:15:26.096295 4677 generic.go:334] "Generic (PLEG): container finished" podID="c2dc4bf3-fecf-4b2a-a664-d5c543196286" containerID="9d309ca1a878a0406de51d6acdf7a0d147f15efbe7ef7d167c8f146d4f3f00ca" exitCode=0 Dec 03 14:15:26 crc kubenswrapper[4677]: I1203 14:15:26.096367 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" event={"ID":"c2dc4bf3-fecf-4b2a-a664-d5c543196286","Type":"ContainerDied","Data":"9d309ca1a878a0406de51d6acdf7a0d147f15efbe7ef7d167c8f146d4f3f00ca"} Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.540911 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.617886 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-repo-setup-combined-ca-bundle\") pod \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.618482 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ktkr\" (UniqueName: \"kubernetes.io/projected/c2dc4bf3-fecf-4b2a-a664-d5c543196286-kube-api-access-5ktkr\") pod \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.618592 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-inventory\") pod \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.623910 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c2dc4bf3-fecf-4b2a-a664-d5c543196286" (UID: "c2dc4bf3-fecf-4b2a-a664-d5c543196286"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.624037 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dc4bf3-fecf-4b2a-a664-d5c543196286-kube-api-access-5ktkr" (OuterVolumeSpecName: "kube-api-access-5ktkr") pod "c2dc4bf3-fecf-4b2a-a664-d5c543196286" (UID: "c2dc4bf3-fecf-4b2a-a664-d5c543196286"). InnerVolumeSpecName "kube-api-access-5ktkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.651778 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-inventory" (OuterVolumeSpecName: "inventory") pod "c2dc4bf3-fecf-4b2a-a664-d5c543196286" (UID: "c2dc4bf3-fecf-4b2a-a664-d5c543196286"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.719629 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-ssh-key\") pod \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\" (UID: \"c2dc4bf3-fecf-4b2a-a664-d5c543196286\") " Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.720097 4677 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.720131 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ktkr\" (UniqueName: \"kubernetes.io/projected/c2dc4bf3-fecf-4b2a-a664-d5c543196286-kube-api-access-5ktkr\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.720149 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.745275 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2dc4bf3-fecf-4b2a-a664-d5c543196286" (UID: "c2dc4bf3-fecf-4b2a-a664-d5c543196286"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:27 crc kubenswrapper[4677]: I1203 14:15:27.822736 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dc4bf3-fecf-4b2a-a664-d5c543196286-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.121758 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" event={"ID":"c2dc4bf3-fecf-4b2a-a664-d5c543196286","Type":"ContainerDied","Data":"7adafff7923900473a3c1c9ffe828b3b4c14abec48e7a46e579594c5e849334a"} Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.121800 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7adafff7923900473a3c1c9ffe828b3b4c14abec48e7a46e579594c5e849334a" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.121912 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.245322 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp"] Dec 03 14:15:28 crc kubenswrapper[4677]: E1203 14:15:28.245815 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dc4bf3-fecf-4b2a-a664-d5c543196286" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.245845 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dc4bf3-fecf-4b2a-a664-d5c543196286" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 14:15:28 crc kubenswrapper[4677]: E1203 14:15:28.245866 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" containerName="collect-profiles" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.245876 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" containerName="collect-profiles" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.246161 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dc4bf3-fecf-4b2a-a664-d5c543196286" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.246195 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" containerName="collect-profiles" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.247045 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.248994 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.252917 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.253005 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.253108 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.257660 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp"] Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.434110 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.434167 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x52dz\" (UniqueName: \"kubernetes.io/projected/c24c4c0c-e7bc-44d1-8f09-99422afc8927-kube-api-access-x52dz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.434245 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.536337 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.536488 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.536512 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x52dz\" (UniqueName: \"kubernetes.io/projected/c24c4c0c-e7bc-44d1-8f09-99422afc8927-kube-api-access-x52dz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.541065 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.541269 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.556926 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x52dz\" (UniqueName: \"kubernetes.io/projected/c24c4c0c-e7bc-44d1-8f09-99422afc8927-kube-api-access-x52dz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8dplp\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.574397 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:28 crc kubenswrapper[4677]: I1203 14:15:28.976378 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:15:28 crc kubenswrapper[4677]: E1203 14:15:28.976999 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:15:29 crc kubenswrapper[4677]: I1203 14:15:29.087243 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp"] Dec 03 14:15:29 crc kubenswrapper[4677]: W1203 14:15:29.090828 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc24c4c0c_e7bc_44d1_8f09_99422afc8927.slice/crio-d6ed57803654d0648db081d4d52b51aa22066005d5b33bc89991e187bfb2ced6 WatchSource:0}: Error finding container d6ed57803654d0648db081d4d52b51aa22066005d5b33bc89991e187bfb2ced6: Status 404 returned error can't find the container with id d6ed57803654d0648db081d4d52b51aa22066005d5b33bc89991e187bfb2ced6 Dec 03 14:15:29 crc kubenswrapper[4677]: I1203 14:15:29.133138 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" event={"ID":"c24c4c0c-e7bc-44d1-8f09-99422afc8927","Type":"ContainerStarted","Data":"d6ed57803654d0648db081d4d52b51aa22066005d5b33bc89991e187bfb2ced6"} Dec 03 14:15:30 crc kubenswrapper[4677]: I1203 14:15:30.144900 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" event={"ID":"c24c4c0c-e7bc-44d1-8f09-99422afc8927","Type":"ContainerStarted","Data":"3fb02802e13e5f6a0bd77c9d6ed355112e5645885b24a62cc0223696f27dd27b"} Dec 03 14:15:30 crc kubenswrapper[4677]: I1203 14:15:30.165774 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" podStartSLOduration=1.73470666 podStartE2EDuration="2.16575179s" podCreationTimestamp="2025-12-03 14:15:28 +0000 UTC" firstStartedPulling="2025-12-03 14:15:29.095047069 +0000 UTC m=+1719.841379524" lastFinishedPulling="2025-12-03 14:15:29.526092199 +0000 UTC m=+1720.272424654" observedRunningTime="2025-12-03 14:15:30.158028615 +0000 UTC m=+1720.904361080" watchObservedRunningTime="2025-12-03 14:15:30.16575179 +0000 UTC m=+1720.912084245" Dec 03 14:15:33 crc kubenswrapper[4677]: I1203 14:15:33.191895 4677 generic.go:334] "Generic (PLEG): container finished" podID="c24c4c0c-e7bc-44d1-8f09-99422afc8927" containerID="3fb02802e13e5f6a0bd77c9d6ed355112e5645885b24a62cc0223696f27dd27b" exitCode=0 Dec 03 14:15:33 crc kubenswrapper[4677]: I1203 14:15:33.192031 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" event={"ID":"c24c4c0c-e7bc-44d1-8f09-99422afc8927","Type":"ContainerDied","Data":"3fb02802e13e5f6a0bd77c9d6ed355112e5645885b24a62cc0223696f27dd27b"} Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.607694 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.758897 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x52dz\" (UniqueName: \"kubernetes.io/projected/c24c4c0c-e7bc-44d1-8f09-99422afc8927-kube-api-access-x52dz\") pod \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.759156 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-ssh-key\") pod \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.759218 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-inventory\") pod \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\" (UID: \"c24c4c0c-e7bc-44d1-8f09-99422afc8927\") " Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.765007 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24c4c0c-e7bc-44d1-8f09-99422afc8927-kube-api-access-x52dz" (OuterVolumeSpecName: "kube-api-access-x52dz") pod "c24c4c0c-e7bc-44d1-8f09-99422afc8927" (UID: "c24c4c0c-e7bc-44d1-8f09-99422afc8927"). InnerVolumeSpecName "kube-api-access-x52dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.791356 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-inventory" (OuterVolumeSpecName: "inventory") pod "c24c4c0c-e7bc-44d1-8f09-99422afc8927" (UID: "c24c4c0c-e7bc-44d1-8f09-99422afc8927"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.791780 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c24c4c0c-e7bc-44d1-8f09-99422afc8927" (UID: "c24c4c0c-e7bc-44d1-8f09-99422afc8927"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.862254 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x52dz\" (UniqueName: \"kubernetes.io/projected/c24c4c0c-e7bc-44d1-8f09-99422afc8927-kube-api-access-x52dz\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.862552 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:34 crc kubenswrapper[4677]: I1203 14:15:34.862628 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c24c4c0c-e7bc-44d1-8f09-99422afc8927-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.219674 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" event={"ID":"c24c4c0c-e7bc-44d1-8f09-99422afc8927","Type":"ContainerDied","Data":"d6ed57803654d0648db081d4d52b51aa22066005d5b33bc89991e187bfb2ced6"} Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.219734 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6ed57803654d0648db081d4d52b51aa22066005d5b33bc89991e187bfb2ced6" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.219794 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8dplp" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.301263 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb"] Dec 03 14:15:35 crc kubenswrapper[4677]: E1203 14:15:35.302062 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24c4c0c-e7bc-44d1-8f09-99422afc8927" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.302170 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24c4c0c-e7bc-44d1-8f09-99422afc8927" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.302606 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24c4c0c-e7bc-44d1-8f09-99422afc8927" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.303680 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.306422 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.306731 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.306884 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.310084 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.320264 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb"] Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.483242 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.483451 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.483492 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z25mq\" (UniqueName: \"kubernetes.io/projected/82c6055b-a376-4795-9d0b-bd03d5b82acd-kube-api-access-z25mq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.483576 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.584572 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.584628 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z25mq\" (UniqueName: \"kubernetes.io/projected/82c6055b-a376-4795-9d0b-bd03d5b82acd-kube-api-access-z25mq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.584718 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.584766 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.591712 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.592003 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.594624 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.604914 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z25mq\" (UniqueName: \"kubernetes.io/projected/82c6055b-a376-4795-9d0b-bd03d5b82acd-kube-api-access-z25mq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:35 crc kubenswrapper[4677]: I1203 14:15:35.626123 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:15:36 crc kubenswrapper[4677]: I1203 14:15:36.148140 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb"] Dec 03 14:15:36 crc kubenswrapper[4677]: I1203 14:15:36.157273 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:15:36 crc kubenswrapper[4677]: I1203 14:15:36.231512 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" event={"ID":"82c6055b-a376-4795-9d0b-bd03d5b82acd","Type":"ContainerStarted","Data":"7598b11d7194034de6817fe213d2fc9589751b4be73c80477378d93c286b48fd"} Dec 03 14:15:39 crc kubenswrapper[4677]: I1203 14:15:39.264691 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" event={"ID":"82c6055b-a376-4795-9d0b-bd03d5b82acd","Type":"ContainerStarted","Data":"01d218fa780e8a984d5a4f0f97f2330f614ce5590f47dd7b57e515b744d554e5"} Dec 03 14:15:39 crc kubenswrapper[4677]: I1203 14:15:39.292357 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" podStartSLOduration=2.339611899 podStartE2EDuration="4.292326991s" podCreationTimestamp="2025-12-03 14:15:35 +0000 UTC" firstStartedPulling="2025-12-03 14:15:36.156999599 +0000 UTC m=+1726.903332044" lastFinishedPulling="2025-12-03 14:15:38.109714671 +0000 UTC m=+1728.856047136" observedRunningTime="2025-12-03 14:15:39.28102713 +0000 UTC m=+1730.027359605" watchObservedRunningTime="2025-12-03 14:15:39.292326991 +0000 UTC m=+1730.038659446" Dec 03 14:15:41 crc kubenswrapper[4677]: I1203 14:15:41.976929 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:15:41 crc kubenswrapper[4677]: E1203 14:15:41.977573 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:15:56 crc kubenswrapper[4677]: I1203 14:15:56.976238 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:15:56 crc kubenswrapper[4677]: E1203 14:15:56.976874 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:16:04 crc kubenswrapper[4677]: I1203 14:16:04.616086 4677 scope.go:117] "RemoveContainer" containerID="1b877a7091dd9c860b5104501ed4c0d071447f1de0bc8d1718d3ab84cfd3e556" Dec 03 14:16:07 crc kubenswrapper[4677]: I1203 14:16:07.977053 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:16:07 crc kubenswrapper[4677]: E1203 14:16:07.977908 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:16:20 crc kubenswrapper[4677]: I1203 14:16:20.976786 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:16:20 crc kubenswrapper[4677]: E1203 14:16:20.977737 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:16:31 crc kubenswrapper[4677]: I1203 14:16:31.976065 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:16:31 crc kubenswrapper[4677]: E1203 14:16:31.977045 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:16:46 crc kubenswrapper[4677]: I1203 14:16:46.975938 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:16:46 crc kubenswrapper[4677]: E1203 14:16:46.976757 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:17:01 crc kubenswrapper[4677]: I1203 14:17:01.978002 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:17:01 crc kubenswrapper[4677]: E1203 14:17:01.979074 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:17:12 crc kubenswrapper[4677]: I1203 14:17:12.976846 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:17:12 crc kubenswrapper[4677]: E1203 14:17:12.978082 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:17:23 crc kubenswrapper[4677]: I1203 14:17:23.977421 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:17:23 crc kubenswrapper[4677]: E1203 14:17:23.978329 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:17:34 crc kubenswrapper[4677]: I1203 14:17:34.976280 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:17:34 crc kubenswrapper[4677]: E1203 14:17:34.977661 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:17:45 crc kubenswrapper[4677]: I1203 14:17:45.976513 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:17:45 crc kubenswrapper[4677]: E1203 14:17:45.977347 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:17:56 crc kubenswrapper[4677]: I1203 14:17:56.976620 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:17:56 crc kubenswrapper[4677]: E1203 14:17:56.977395 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:18:10 crc kubenswrapper[4677]: I1203 14:18:10.976442 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:18:10 crc kubenswrapper[4677]: E1203 14:18:10.978554 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:18:21 crc kubenswrapper[4677]: I1203 14:18:21.976074 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:18:21 crc kubenswrapper[4677]: E1203 14:18:21.976842 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:18:36 crc kubenswrapper[4677]: I1203 14:18:36.976923 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:18:36 crc kubenswrapper[4677]: E1203 14:18:36.977762 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:18:47 crc kubenswrapper[4677]: I1203 14:18:47.977170 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:18:48 crc kubenswrapper[4677]: I1203 14:18:48.354853 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"dc95799bd3832da5bee136d530caa6131c382dcae6b9ac4b24eb38e3ea0e1e21"} Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.087642 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-da07-account-create-update-fpsng"] Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.096170 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ztpkm"] Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.104833 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bbef-account-create-update-8btps"] Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.113340 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-zk8qb"] Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.122244 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bbef-account-create-update-8btps"] Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.130532 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-zk8qb"] Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.138409 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ztpkm"] Dec 03 14:18:58 crc kubenswrapper[4677]: I1203 14:18:58.146007 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-da07-account-create-update-fpsng"] Dec 03 14:18:59 crc kubenswrapper[4677]: I1203 14:18:59.986675 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2994ba2a-3d45-4303-b0c1-9800979e1cb5" path="/var/lib/kubelet/pods/2994ba2a-3d45-4303-b0c1-9800979e1cb5/volumes" Dec 03 14:18:59 crc kubenswrapper[4677]: I1203 14:18:59.988007 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445b8de4-6203-494a-9448-86f12765390e" path="/var/lib/kubelet/pods/445b8de4-6203-494a-9448-86f12765390e/volumes" Dec 03 14:18:59 crc kubenswrapper[4677]: I1203 14:18:59.988708 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0" path="/var/lib/kubelet/pods/5c836fd0-4bf6-4a43-b62b-6ad5d12f3bd0/volumes" Dec 03 14:18:59 crc kubenswrapper[4677]: I1203 14:18:59.989451 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7c96896-06fc-4875-a6c0-f0f3cec43b90" path="/var/lib/kubelet/pods/e7c96896-06fc-4875-a6c0-f0f3cec43b90/volumes" Dec 03 14:19:01 crc kubenswrapper[4677]: I1203 14:19:01.036687 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-rh642"] Dec 03 14:19:01 crc kubenswrapper[4677]: I1203 14:19:01.049173 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-6844-account-create-update-t6r4j"] Dec 03 14:19:01 crc kubenswrapper[4677]: I1203 14:19:01.059722 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-6844-account-create-update-t6r4j"] Dec 03 14:19:01 crc kubenswrapper[4677]: I1203 14:19:01.072135 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-rh642"] Dec 03 14:19:01 crc kubenswrapper[4677]: I1203 14:19:01.987662 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1" path="/var/lib/kubelet/pods/67a0d4af-5eb2-42e1-b66d-a4f9d5e4c1c1/volumes" Dec 03 14:19:01 crc kubenswrapper[4677]: I1203 14:19:01.988389 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf253b5-7e86-4dfe-9b49-5fa2869b2473" path="/var/lib/kubelet/pods/7cf253b5-7e86-4dfe-9b49-5fa2869b2473/volumes" Dec 03 14:19:04 crc kubenswrapper[4677]: I1203 14:19:04.731918 4677 scope.go:117] "RemoveContainer" containerID="6799fc13216f7781718fd67889d0ea610dc281c4276d9b182d12d7c03c456825" Dec 03 14:19:04 crc kubenswrapper[4677]: I1203 14:19:04.762025 4677 scope.go:117] "RemoveContainer" containerID="4a61e61e88202d5376b223531762977cf7f45a57bfcca8dc7971bc4250fa2687" Dec 03 14:19:04 crc kubenswrapper[4677]: I1203 14:19:04.825916 4677 scope.go:117] "RemoveContainer" containerID="dbf149195267d8880ff1c49682426d35043703a40b9bf8166608aba2f0993b46" Dec 03 14:19:04 crc kubenswrapper[4677]: I1203 14:19:04.859916 4677 scope.go:117] "RemoveContainer" containerID="07bc48d960c258535c88f43adf264d48d0a2a004d9d835a7e8edabbd19a33f8f" Dec 03 14:19:04 crc kubenswrapper[4677]: I1203 14:19:04.895421 4677 scope.go:117] "RemoveContainer" containerID="44a8375ab0ed916c8d8c90009b485a19afb22f436de367ac58b3af7b022fb996" Dec 03 14:19:04 crc kubenswrapper[4677]: I1203 14:19:04.942198 4677 scope.go:117] "RemoveContainer" containerID="4d6bfbd19412d37509ea4d48bf3b92a4f7ef6bdbf2a405ea46c9ca9b55f2aa48" Dec 03 14:19:04 crc kubenswrapper[4677]: I1203 14:19:04.996048 4677 scope.go:117] "RemoveContainer" containerID="bd4cf25b1ae6dbc951dabe951610ae35ea15903d469d83a681b551add1ff2a2f" Dec 03 14:19:05 crc kubenswrapper[4677]: I1203 14:19:05.044592 4677 scope.go:117] "RemoveContainer" containerID="360d3c2f92c2bba064ddada31151b6307864efb3175f5e930fe2206dd1b9918a" Dec 03 14:19:05 crc kubenswrapper[4677]: I1203 14:19:05.064265 4677 scope.go:117] "RemoveContainer" containerID="a1a1b69dfb2bba22c132554d8465f8181d225f032961bf190cd36b8d1ed3b8b2" Dec 03 14:19:20 crc kubenswrapper[4677]: I1203 14:19:20.667670 4677 generic.go:334] "Generic (PLEG): container finished" podID="82c6055b-a376-4795-9d0b-bd03d5b82acd" containerID="01d218fa780e8a984d5a4f0f97f2330f614ce5590f47dd7b57e515b744d554e5" exitCode=0 Dec 03 14:19:20 crc kubenswrapper[4677]: I1203 14:19:20.667757 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" event={"ID":"82c6055b-a376-4795-9d0b-bd03d5b82acd","Type":"ContainerDied","Data":"01d218fa780e8a984d5a4f0f97f2330f614ce5590f47dd7b57e515b744d554e5"} Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.109662 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.221131 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-bootstrap-combined-ca-bundle\") pod \"82c6055b-a376-4795-9d0b-bd03d5b82acd\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.221276 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z25mq\" (UniqueName: \"kubernetes.io/projected/82c6055b-a376-4795-9d0b-bd03d5b82acd-kube-api-access-z25mq\") pod \"82c6055b-a376-4795-9d0b-bd03d5b82acd\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.221537 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-ssh-key\") pod \"82c6055b-a376-4795-9d0b-bd03d5b82acd\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.221569 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-inventory\") pod \"82c6055b-a376-4795-9d0b-bd03d5b82acd\" (UID: \"82c6055b-a376-4795-9d0b-bd03d5b82acd\") " Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.233748 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "82c6055b-a376-4795-9d0b-bd03d5b82acd" (UID: "82c6055b-a376-4795-9d0b-bd03d5b82acd"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.233867 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c6055b-a376-4795-9d0b-bd03d5b82acd-kube-api-access-z25mq" (OuterVolumeSpecName: "kube-api-access-z25mq") pod "82c6055b-a376-4795-9d0b-bd03d5b82acd" (UID: "82c6055b-a376-4795-9d0b-bd03d5b82acd"). InnerVolumeSpecName "kube-api-access-z25mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.257765 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-inventory" (OuterVolumeSpecName: "inventory") pod "82c6055b-a376-4795-9d0b-bd03d5b82acd" (UID: "82c6055b-a376-4795-9d0b-bd03d5b82acd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.262125 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "82c6055b-a376-4795-9d0b-bd03d5b82acd" (UID: "82c6055b-a376-4795-9d0b-bd03d5b82acd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.324076 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.324137 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.324153 4677 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c6055b-a376-4795-9d0b-bd03d5b82acd-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.324178 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z25mq\" (UniqueName: \"kubernetes.io/projected/82c6055b-a376-4795-9d0b-bd03d5b82acd-kube-api-access-z25mq\") on node \"crc\" DevicePath \"\"" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.687624 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" event={"ID":"82c6055b-a376-4795-9d0b-bd03d5b82acd","Type":"ContainerDied","Data":"7598b11d7194034de6817fe213d2fc9589751b4be73c80477378d93c286b48fd"} Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.687965 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7598b11d7194034de6817fe213d2fc9589751b4be73c80477378d93c286b48fd" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.688041 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.797284 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q"] Dec 03 14:19:22 crc kubenswrapper[4677]: E1203 14:19:22.797810 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c6055b-a376-4795-9d0b-bd03d5b82acd" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.797832 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c6055b-a376-4795-9d0b-bd03d5b82acd" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.798147 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c6055b-a376-4795-9d0b-bd03d5b82acd" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.799014 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.801688 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.802158 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.802260 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.808073 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.812304 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q"] Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.936743 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmz7m\" (UniqueName: \"kubernetes.io/projected/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-kube-api-access-hmz7m\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.936968 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:22 crc kubenswrapper[4677]: I1203 14:19:22.937069 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.038729 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmz7m\" (UniqueName: \"kubernetes.io/projected/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-kube-api-access-hmz7m\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.038830 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.038874 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.045748 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.046031 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.060871 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmz7m\" (UniqueName: \"kubernetes.io/projected/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-kube-api-access-hmz7m\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-25g8q\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.130325 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.672684 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q"] Dec 03 14:19:23 crc kubenswrapper[4677]: I1203 14:19:23.712449 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" event={"ID":"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737","Type":"ContainerStarted","Data":"495e72f71d1eefb84cb827b61cee7409142a9414f416e1ce7f0dcef7385389d2"} Dec 03 14:19:24 crc kubenswrapper[4677]: I1203 14:19:24.722748 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" event={"ID":"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737","Type":"ContainerStarted","Data":"1bb3cab2f91e90c85f97b83321a8b589885c42b1c6504c6c9e839f91a665a3ee"} Dec 03 14:19:24 crc kubenswrapper[4677]: I1203 14:19:24.747831 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" podStartSLOduration=2.215505854 podStartE2EDuration="2.74780716s" podCreationTimestamp="2025-12-03 14:19:22 +0000 UTC" firstStartedPulling="2025-12-03 14:19:23.678418224 +0000 UTC m=+1954.424750679" lastFinishedPulling="2025-12-03 14:19:24.21071954 +0000 UTC m=+1954.957051985" observedRunningTime="2025-12-03 14:19:24.738740069 +0000 UTC m=+1955.485072544" watchObservedRunningTime="2025-12-03 14:19:24.74780716 +0000 UTC m=+1955.494139625" Dec 03 14:19:32 crc kubenswrapper[4677]: I1203 14:19:32.055810 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ncgkf"] Dec 03 14:19:32 crc kubenswrapper[4677]: I1203 14:19:32.064297 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6be3-account-create-update-pj4dm"] Dec 03 14:19:32 crc kubenswrapper[4677]: I1203 14:19:32.072031 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ncgkf"] Dec 03 14:19:32 crc kubenswrapper[4677]: I1203 14:19:32.079216 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6be3-account-create-update-pj4dm"] Dec 03 14:19:33 crc kubenswrapper[4677]: I1203 14:19:33.992448 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0942d3aa-be7a-47e1-a5a5-7fcae1993259" path="/var/lib/kubelet/pods/0942d3aa-be7a-47e1-a5a5-7fcae1993259/volumes" Dec 03 14:19:33 crc kubenswrapper[4677]: I1203 14:19:33.993595 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8853a43-b0b5-46da-932d-53826ee8cd4e" path="/var/lib/kubelet/pods/a8853a43-b0b5-46da-932d-53826ee8cd4e/volumes" Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.036263 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-r2jg2"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.050872 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-94ctf"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.063706 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-94ctf"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.075464 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-cdnh4"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.089619 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-r2jg2"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.102141 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-2f2f-account-create-update-wvblx"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.111967 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-cf68-account-create-update-wppvc"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.129630 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-cdnh4"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.144228 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-2f2f-account-create-update-wvblx"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.156037 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-cf68-account-create-update-wppvc"] Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.989146 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20bee223-5339-4fe5-a341-9bd1b1904079" path="/var/lib/kubelet/pods/20bee223-5339-4fe5-a341-9bd1b1904079/volumes" Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.990563 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d278aed-ed7e-49fe-b952-899f9d0076ef" path="/var/lib/kubelet/pods/7d278aed-ed7e-49fe-b952-899f9d0076ef/volumes" Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.991452 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d336d821-9f68-4486-95f7-e8c79c65a928" path="/var/lib/kubelet/pods/d336d821-9f68-4486-95f7-e8c79c65a928/volumes" Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.992333 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dded6789-7bf2-4fd7-8d73-6a11c4ad48d3" path="/var/lib/kubelet/pods/dded6789-7bf2-4fd7-8d73-6a11c4ad48d3/volumes" Dec 03 14:19:37 crc kubenswrapper[4677]: I1203 14:19:37.993773 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea30c0a1-61fd-45e9-905b-0f022a6381bd" path="/var/lib/kubelet/pods/ea30c0a1-61fd-45e9-905b-0f022a6381bd/volumes" Dec 03 14:19:43 crc kubenswrapper[4677]: I1203 14:19:43.034288 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6850-account-create-update-8lrt8"] Dec 03 14:19:43 crc kubenswrapper[4677]: I1203 14:19:43.044584 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6850-account-create-update-8lrt8"] Dec 03 14:19:43 crc kubenswrapper[4677]: I1203 14:19:43.990309 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b91633d9-71f9-40bc-832d-b24143d8478b" path="/var/lib/kubelet/pods/b91633d9-71f9-40bc-832d-b24143d8478b/volumes" Dec 03 14:19:56 crc kubenswrapper[4677]: I1203 14:19:56.048389 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-94ppc"] Dec 03 14:19:56 crc kubenswrapper[4677]: I1203 14:19:56.058193 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-r9jq5"] Dec 03 14:19:56 crc kubenswrapper[4677]: I1203 14:19:56.067749 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-r9jq5"] Dec 03 14:19:56 crc kubenswrapper[4677]: I1203 14:19:56.078541 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-94ppc"] Dec 03 14:19:57 crc kubenswrapper[4677]: I1203 14:19:57.986709 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aaf4f81-d89b-40d9-af15-4920cdf73a76" path="/var/lib/kubelet/pods/3aaf4f81-d89b-40d9-af15-4920cdf73a76/volumes" Dec 03 14:19:57 crc kubenswrapper[4677]: I1203 14:19:57.987488 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="711edb9c-bc87-4f0a-adea-36ef2502783f" path="/var/lib/kubelet/pods/711edb9c-bc87-4f0a-adea-36ef2502783f/volumes" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.217600 4677 scope.go:117] "RemoveContainer" containerID="40f7a4ad7c5d67b5338d8fed11de0ff8501f0df2a42f5b4cefe0f762f8082d8c" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.262015 4677 scope.go:117] "RemoveContainer" containerID="6bb0a2c1588e6f8de958dc087c06cef727b1c065aca436183fd2cda4fdd92192" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.305331 4677 scope.go:117] "RemoveContainer" containerID="eddaa4406812d8cc571e839129858025034f68c41b6c802b490846d25e716fb3" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.352493 4677 scope.go:117] "RemoveContainer" containerID="eeb3279669388080e59270473bb3fd077f6f32ab181915561302cd7fab2ddd61" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.395367 4677 scope.go:117] "RemoveContainer" containerID="ff98e5e99d608250b4255d7926137dc9bb4b9476c5a6305d6e3a01940a37a6a2" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.444156 4677 scope.go:117] "RemoveContainer" containerID="b55fe36f06bde617e131cb67859b8773719c686e8ea20b017582dd9a095de032" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.511892 4677 scope.go:117] "RemoveContainer" containerID="31c5780adee53a71145c6e49c64f0976ad29c84c1f9dbdc46098f1666a07d86f" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.532602 4677 scope.go:117] "RemoveContainer" containerID="49cc87b587f88610d2dd11b90c14ba03581c6f96ecdc90e9aaca9eb7d6623e48" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.552479 4677 scope.go:117] "RemoveContainer" containerID="9b976fe0d50cabd937b9fccfef33cda3f678185496529614cd2d77251bc04f72" Dec 03 14:20:05 crc kubenswrapper[4677]: I1203 14:20:05.574717 4677 scope.go:117] "RemoveContainer" containerID="4b88e476aaf26876be26a3ce964bed0b3f0751c4207052c1d5986fbfdc7b3433" Dec 03 14:20:32 crc kubenswrapper[4677]: I1203 14:20:32.061671 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-kbkxk"] Dec 03 14:20:32 crc kubenswrapper[4677]: I1203 14:20:32.073014 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-kbkxk"] Dec 03 14:20:33 crc kubenswrapper[4677]: I1203 14:20:33.994694 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c12c954-d94a-404a-9263-a8371fdc12e4" path="/var/lib/kubelet/pods/4c12c954-d94a-404a-9263-a8371fdc12e4/volumes" Dec 03 14:20:34 crc kubenswrapper[4677]: I1203 14:20:34.038203 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-72sks"] Dec 03 14:20:34 crc kubenswrapper[4677]: I1203 14:20:34.051863 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-72sks"] Dec 03 14:20:35 crc kubenswrapper[4677]: I1203 14:20:35.987842 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d88b54f3-fe9e-4247-8292-1f513f84fd33" path="/var/lib/kubelet/pods/d88b54f3-fe9e-4247-8292-1f513f84fd33/volumes" Dec 03 14:20:48 crc kubenswrapper[4677]: I1203 14:20:48.034602 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-f74nz"] Dec 03 14:20:48 crc kubenswrapper[4677]: I1203 14:20:48.045555 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ff56g"] Dec 03 14:20:48 crc kubenswrapper[4677]: I1203 14:20:48.055308 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-f74nz"] Dec 03 14:20:48 crc kubenswrapper[4677]: I1203 14:20:48.063858 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ff56g"] Dec 03 14:20:49 crc kubenswrapper[4677]: I1203 14:20:49.994691 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24ec7657-9bc6-4953-9b8c-d07a62b7aff1" path="/var/lib/kubelet/pods/24ec7657-9bc6-4953-9b8c-d07a62b7aff1/volumes" Dec 03 14:20:49 crc kubenswrapper[4677]: I1203 14:20:49.997661 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47d54aae-4e2b-454f-a29d-9b252fd60179" path="/var/lib/kubelet/pods/47d54aae-4e2b-454f-a29d-9b252fd60179/volumes" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.579419 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jgpns"] Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.582888 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.597774 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jgpns"] Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.720293 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wp56\" (UniqueName: \"kubernetes.io/projected/13baedf7-2496-48d4-b165-1efddc4e20a0-kube-api-access-7wp56\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.720660 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-utilities\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.721443 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-catalog-content\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.823126 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-catalog-content\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.823226 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wp56\" (UniqueName: \"kubernetes.io/projected/13baedf7-2496-48d4-b165-1efddc4e20a0-kube-api-access-7wp56\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.823349 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-utilities\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.823664 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-catalog-content\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.824019 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-utilities\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.846299 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wp56\" (UniqueName: \"kubernetes.io/projected/13baedf7-2496-48d4-b165-1efddc4e20a0-kube-api-access-7wp56\") pod \"community-operators-jgpns\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:56 crc kubenswrapper[4677]: I1203 14:20:56.909970 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:20:57 crc kubenswrapper[4677]: I1203 14:20:57.440256 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jgpns"] Dec 03 14:20:57 crc kubenswrapper[4677]: I1203 14:20:57.657299 4677 generic.go:334] "Generic (PLEG): container finished" podID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerID="a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff" exitCode=0 Dec 03 14:20:57 crc kubenswrapper[4677]: I1203 14:20:57.657464 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgpns" event={"ID":"13baedf7-2496-48d4-b165-1efddc4e20a0","Type":"ContainerDied","Data":"a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff"} Dec 03 14:20:57 crc kubenswrapper[4677]: I1203 14:20:57.657637 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgpns" event={"ID":"13baedf7-2496-48d4-b165-1efddc4e20a0","Type":"ContainerStarted","Data":"f0a521dd5673cea2c99386a276fad0624c33d59aa6e64f9f661fa7503f59a207"} Dec 03 14:20:57 crc kubenswrapper[4677]: I1203 14:20:57.659620 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.370721 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-92lbj"] Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.373223 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.392625 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-92lbj"] Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.468076 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-catalog-content\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.468498 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctb4t\" (UniqueName: \"kubernetes.io/projected/b57e4890-c847-427c-a4cc-b11bab6891c2-kube-api-access-ctb4t\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.468551 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-utilities\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.570899 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctb4t\" (UniqueName: \"kubernetes.io/projected/b57e4890-c847-427c-a4cc-b11bab6891c2-kube-api-access-ctb4t\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.570989 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-utilities\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.571233 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-catalog-content\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.571528 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-utilities\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.571725 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-catalog-content\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.589550 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctb4t\" (UniqueName: \"kubernetes.io/projected/b57e4890-c847-427c-a4cc-b11bab6891c2-kube-api-access-ctb4t\") pod \"redhat-marketplace-92lbj\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:58 crc kubenswrapper[4677]: I1203 14:20:58.700863 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:20:59 crc kubenswrapper[4677]: I1203 14:20:59.167718 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-92lbj"] Dec 03 14:20:59 crc kubenswrapper[4677]: W1203 14:20:59.171093 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb57e4890_c847_427c_a4cc_b11bab6891c2.slice/crio-9f26499c378511c3dd7a4b12f5dc41cf8072ab94838d6784cfcd5e3e91902b77 WatchSource:0}: Error finding container 9f26499c378511c3dd7a4b12f5dc41cf8072ab94838d6784cfcd5e3e91902b77: Status 404 returned error can't find the container with id 9f26499c378511c3dd7a4b12f5dc41cf8072ab94838d6784cfcd5e3e91902b77 Dec 03 14:20:59 crc kubenswrapper[4677]: I1203 14:20:59.681008 4677 generic.go:334] "Generic (PLEG): container finished" podID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerID="e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966" exitCode=0 Dec 03 14:20:59 crc kubenswrapper[4677]: I1203 14:20:59.681085 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-92lbj" event={"ID":"b57e4890-c847-427c-a4cc-b11bab6891c2","Type":"ContainerDied","Data":"e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966"} Dec 03 14:20:59 crc kubenswrapper[4677]: I1203 14:20:59.681400 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-92lbj" event={"ID":"b57e4890-c847-427c-a4cc-b11bab6891c2","Type":"ContainerStarted","Data":"9f26499c378511c3dd7a4b12f5dc41cf8072ab94838d6784cfcd5e3e91902b77"} Dec 03 14:20:59 crc kubenswrapper[4677]: I1203 14:20:59.688064 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgpns" event={"ID":"13baedf7-2496-48d4-b165-1efddc4e20a0","Type":"ContainerStarted","Data":"b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10"} Dec 03 14:21:00 crc kubenswrapper[4677]: I1203 14:21:00.042149 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-njl82"] Dec 03 14:21:00 crc kubenswrapper[4677]: I1203 14:21:00.050431 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-njl82"] Dec 03 14:21:00 crc kubenswrapper[4677]: I1203 14:21:00.703449 4677 generic.go:334] "Generic (PLEG): container finished" podID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerID="b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10" exitCode=0 Dec 03 14:21:00 crc kubenswrapper[4677]: I1203 14:21:00.703512 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgpns" event={"ID":"13baedf7-2496-48d4-b165-1efddc4e20a0","Type":"ContainerDied","Data":"b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10"} Dec 03 14:21:01 crc kubenswrapper[4677]: I1203 14:21:01.031307 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-hgx4j"] Dec 03 14:21:01 crc kubenswrapper[4677]: I1203 14:21:01.042245 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-hgx4j"] Dec 03 14:21:01 crc kubenswrapper[4677]: I1203 14:21:01.987248 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fd09b0c-1690-42e1-b97a-a18e7fefc813" path="/var/lib/kubelet/pods/0fd09b0c-1690-42e1-b97a-a18e7fefc813/volumes" Dec 03 14:21:01 crc kubenswrapper[4677]: I1203 14:21:01.988178 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25625b10-c45e-40f0-b342-bdcdc3e79c9c" path="/var/lib/kubelet/pods/25625b10-c45e-40f0-b342-bdcdc3e79c9c/volumes" Dec 03 14:21:02 crc kubenswrapper[4677]: I1203 14:21:02.723870 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-92lbj" event={"ID":"b57e4890-c847-427c-a4cc-b11bab6891c2","Type":"ContainerStarted","Data":"57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07"} Dec 03 14:21:02 crc kubenswrapper[4677]: I1203 14:21:02.726685 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgpns" event={"ID":"13baedf7-2496-48d4-b165-1efddc4e20a0","Type":"ContainerStarted","Data":"376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b"} Dec 03 14:21:02 crc kubenswrapper[4677]: I1203 14:21:02.776719 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jgpns" podStartSLOduration=2.365813773 podStartE2EDuration="6.77670126s" podCreationTimestamp="2025-12-03 14:20:56 +0000 UTC" firstStartedPulling="2025-12-03 14:20:57.659394213 +0000 UTC m=+2048.405726658" lastFinishedPulling="2025-12-03 14:21:02.07028169 +0000 UTC m=+2052.816614145" observedRunningTime="2025-12-03 14:21:02.773177742 +0000 UTC m=+2053.519510207" watchObservedRunningTime="2025-12-03 14:21:02.77670126 +0000 UTC m=+2053.523033715" Dec 03 14:21:03 crc kubenswrapper[4677]: I1203 14:21:03.744137 4677 generic.go:334] "Generic (PLEG): container finished" podID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerID="57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07" exitCode=0 Dec 03 14:21:03 crc kubenswrapper[4677]: I1203 14:21:03.744371 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-92lbj" event={"ID":"b57e4890-c847-427c-a4cc-b11bab6891c2","Type":"ContainerDied","Data":"57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07"} Dec 03 14:21:04 crc kubenswrapper[4677]: I1203 14:21:04.760100 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-92lbj" event={"ID":"b57e4890-c847-427c-a4cc-b11bab6891c2","Type":"ContainerStarted","Data":"042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547"} Dec 03 14:21:04 crc kubenswrapper[4677]: I1203 14:21:04.784417 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-92lbj" podStartSLOduration=2.29398632 podStartE2EDuration="6.78439928s" podCreationTimestamp="2025-12-03 14:20:58 +0000 UTC" firstStartedPulling="2025-12-03 14:20:59.682617763 +0000 UTC m=+2050.428950218" lastFinishedPulling="2025-12-03 14:21:04.173030723 +0000 UTC m=+2054.919363178" observedRunningTime="2025-12-03 14:21:04.779564597 +0000 UTC m=+2055.525897062" watchObservedRunningTime="2025-12-03 14:21:04.78439928 +0000 UTC m=+2055.530731735" Dec 03 14:21:05 crc kubenswrapper[4677]: I1203 14:21:05.759027 4677 scope.go:117] "RemoveContainer" containerID="acb134cff354677e06caeafe5b47abfe39f4ca0e2990bea9a3e8de61c3fc2a5e" Dec 03 14:21:05 crc kubenswrapper[4677]: I1203 14:21:05.793960 4677 scope.go:117] "RemoveContainer" containerID="12faf5c4818df4effa9de71c604ec4af267f4cd5b4a81c1535a6a7c5920ed641" Dec 03 14:21:05 crc kubenswrapper[4677]: I1203 14:21:05.841436 4677 scope.go:117] "RemoveContainer" containerID="8c5473373d1e776782a3579f5497a7aee84971d76b30289c6cc605d85f68a020" Dec 03 14:21:05 crc kubenswrapper[4677]: I1203 14:21:05.892896 4677 scope.go:117] "RemoveContainer" containerID="14b20e13076483ca1cbb724b3434ea4e9211378877f0effb2a1ece94cb676338" Dec 03 14:21:05 crc kubenswrapper[4677]: I1203 14:21:05.945272 4677 scope.go:117] "RemoveContainer" containerID="bfe141487a2b8cfff701b2ecbb5a27bcd223a4147c90788060d261ba646c6a83" Dec 03 14:21:05 crc kubenswrapper[4677]: I1203 14:21:05.995741 4677 scope.go:117] "RemoveContainer" containerID="64425cc924e22eaa6e7d652322613611f220ab3d76f8ce2539e1804f9ab5321a" Dec 03 14:21:06 crc kubenswrapper[4677]: I1203 14:21:06.910768 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:21:06 crc kubenswrapper[4677]: I1203 14:21:06.911249 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:21:06 crc kubenswrapper[4677]: I1203 14:21:06.958639 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:21:07 crc kubenswrapper[4677]: I1203 14:21:07.838960 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:21:08 crc kubenswrapper[4677]: I1203 14:21:08.161586 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jgpns"] Dec 03 14:21:08 crc kubenswrapper[4677]: I1203 14:21:08.437724 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:21:08 crc kubenswrapper[4677]: I1203 14:21:08.437820 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:21:08 crc kubenswrapper[4677]: I1203 14:21:08.701429 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:21:08 crc kubenswrapper[4677]: I1203 14:21:08.701496 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:21:08 crc kubenswrapper[4677]: I1203 14:21:08.770804 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:21:09 crc kubenswrapper[4677]: I1203 14:21:09.819216 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jgpns" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="registry-server" containerID="cri-o://376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b" gracePeriod=2 Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.296502 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.422349 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-catalog-content\") pod \"13baedf7-2496-48d4-b165-1efddc4e20a0\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.422489 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-utilities\") pod \"13baedf7-2496-48d4-b165-1efddc4e20a0\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.422668 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wp56\" (UniqueName: \"kubernetes.io/projected/13baedf7-2496-48d4-b165-1efddc4e20a0-kube-api-access-7wp56\") pod \"13baedf7-2496-48d4-b165-1efddc4e20a0\" (UID: \"13baedf7-2496-48d4-b165-1efddc4e20a0\") " Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.423063 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-utilities" (OuterVolumeSpecName: "utilities") pod "13baedf7-2496-48d4-b165-1efddc4e20a0" (UID: "13baedf7-2496-48d4-b165-1efddc4e20a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.423215 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.431665 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13baedf7-2496-48d4-b165-1efddc4e20a0-kube-api-access-7wp56" (OuterVolumeSpecName: "kube-api-access-7wp56") pod "13baedf7-2496-48d4-b165-1efddc4e20a0" (UID: "13baedf7-2496-48d4-b165-1efddc4e20a0"). InnerVolumeSpecName "kube-api-access-7wp56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.476343 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13baedf7-2496-48d4-b165-1efddc4e20a0" (UID: "13baedf7-2496-48d4-b165-1efddc4e20a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.525405 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wp56\" (UniqueName: \"kubernetes.io/projected/13baedf7-2496-48d4-b165-1efddc4e20a0-kube-api-access-7wp56\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.542671 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13baedf7-2496-48d4-b165-1efddc4e20a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.829815 4677 generic.go:334] "Generic (PLEG): container finished" podID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerID="376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b" exitCode=0 Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.829866 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgpns" event={"ID":"13baedf7-2496-48d4-b165-1efddc4e20a0","Type":"ContainerDied","Data":"376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b"} Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.829896 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgpns" event={"ID":"13baedf7-2496-48d4-b165-1efddc4e20a0","Type":"ContainerDied","Data":"f0a521dd5673cea2c99386a276fad0624c33d59aa6e64f9f661fa7503f59a207"} Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.829920 4677 scope.go:117] "RemoveContainer" containerID="376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.831377 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgpns" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.850977 4677 scope.go:117] "RemoveContainer" containerID="b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.878177 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jgpns"] Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.887742 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jgpns"] Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.893724 4677 scope.go:117] "RemoveContainer" containerID="a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.927133 4677 scope.go:117] "RemoveContainer" containerID="376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b" Dec 03 14:21:10 crc kubenswrapper[4677]: E1203 14:21:10.927991 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b\": container with ID starting with 376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b not found: ID does not exist" containerID="376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.928042 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b"} err="failed to get container status \"376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b\": rpc error: code = NotFound desc = could not find container \"376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b\": container with ID starting with 376f85c9f2c750a12ea0d7ca298b027ae73f6c790cce7380a7eb54a359a4762b not found: ID does not exist" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.928069 4677 scope.go:117] "RemoveContainer" containerID="b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10" Dec 03 14:21:10 crc kubenswrapper[4677]: E1203 14:21:10.928391 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10\": container with ID starting with b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10 not found: ID does not exist" containerID="b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.928431 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10"} err="failed to get container status \"b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10\": rpc error: code = NotFound desc = could not find container \"b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10\": container with ID starting with b9814b57ca03297756a0f369bbc6a92cd55c92d0e0615957a014a573959a1c10 not found: ID does not exist" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.928452 4677 scope.go:117] "RemoveContainer" containerID="a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff" Dec 03 14:21:10 crc kubenswrapper[4677]: E1203 14:21:10.928851 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff\": container with ID starting with a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff not found: ID does not exist" containerID="a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff" Dec 03 14:21:10 crc kubenswrapper[4677]: I1203 14:21:10.928918 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff"} err="failed to get container status \"a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff\": rpc error: code = NotFound desc = could not find container \"a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff\": container with ID starting with a49cfb22d13655e6a3dadeafd6ae108af86f79fa04c99a337f0ee3513f12aaff not found: ID does not exist" Dec 03 14:21:11 crc kubenswrapper[4677]: I1203 14:21:11.991133 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" path="/var/lib/kubelet/pods/13baedf7-2496-48d4-b165-1efddc4e20a0/volumes" Dec 03 14:21:18 crc kubenswrapper[4677]: I1203 14:21:18.756429 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:21:18 crc kubenswrapper[4677]: I1203 14:21:18.813102 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-92lbj"] Dec 03 14:21:18 crc kubenswrapper[4677]: I1203 14:21:18.901753 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-92lbj" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="registry-server" containerID="cri-o://042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547" gracePeriod=2 Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.342629 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.521510 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-utilities\") pod \"b57e4890-c847-427c-a4cc-b11bab6891c2\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.521597 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctb4t\" (UniqueName: \"kubernetes.io/projected/b57e4890-c847-427c-a4cc-b11bab6891c2-kube-api-access-ctb4t\") pod \"b57e4890-c847-427c-a4cc-b11bab6891c2\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.521645 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-catalog-content\") pod \"b57e4890-c847-427c-a4cc-b11bab6891c2\" (UID: \"b57e4890-c847-427c-a4cc-b11bab6891c2\") " Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.529421 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-utilities" (OuterVolumeSpecName: "utilities") pod "b57e4890-c847-427c-a4cc-b11bab6891c2" (UID: "b57e4890-c847-427c-a4cc-b11bab6891c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.535429 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b57e4890-c847-427c-a4cc-b11bab6891c2-kube-api-access-ctb4t" (OuterVolumeSpecName: "kube-api-access-ctb4t") pod "b57e4890-c847-427c-a4cc-b11bab6891c2" (UID: "b57e4890-c847-427c-a4cc-b11bab6891c2"). InnerVolumeSpecName "kube-api-access-ctb4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.547705 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b57e4890-c847-427c-a4cc-b11bab6891c2" (UID: "b57e4890-c847-427c-a4cc-b11bab6891c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.624344 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.624408 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctb4t\" (UniqueName: \"kubernetes.io/projected/b57e4890-c847-427c-a4cc-b11bab6891c2-kube-api-access-ctb4t\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.624427 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57e4890-c847-427c-a4cc-b11bab6891c2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.915341 4677 generic.go:334] "Generic (PLEG): container finished" podID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerID="042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547" exitCode=0 Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.915428 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-92lbj" event={"ID":"b57e4890-c847-427c-a4cc-b11bab6891c2","Type":"ContainerDied","Data":"042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547"} Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.915497 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-92lbj" event={"ID":"b57e4890-c847-427c-a4cc-b11bab6891c2","Type":"ContainerDied","Data":"9f26499c378511c3dd7a4b12f5dc41cf8072ab94838d6784cfcd5e3e91902b77"} Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.915524 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-92lbj" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.915538 4677 scope.go:117] "RemoveContainer" containerID="042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.950375 4677 scope.go:117] "RemoveContainer" containerID="57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.971515 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-92lbj"] Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.994496 4677 scope.go:117] "RemoveContainer" containerID="e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966" Dec 03 14:21:19 crc kubenswrapper[4677]: I1203 14:21:19.995264 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-92lbj"] Dec 03 14:21:20 crc kubenswrapper[4677]: I1203 14:21:20.029323 4677 scope.go:117] "RemoveContainer" containerID="042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547" Dec 03 14:21:20 crc kubenswrapper[4677]: E1203 14:21:20.029878 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547\": container with ID starting with 042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547 not found: ID does not exist" containerID="042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547" Dec 03 14:21:20 crc kubenswrapper[4677]: I1203 14:21:20.029914 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547"} err="failed to get container status \"042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547\": rpc error: code = NotFound desc = could not find container \"042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547\": container with ID starting with 042e3c9a861bbad8e275052a81724ad25f8a17904d3304ff4b0dc8f415840547 not found: ID does not exist" Dec 03 14:21:20 crc kubenswrapper[4677]: I1203 14:21:20.029934 4677 scope.go:117] "RemoveContainer" containerID="57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07" Dec 03 14:21:20 crc kubenswrapper[4677]: E1203 14:21:20.030230 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07\": container with ID starting with 57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07 not found: ID does not exist" containerID="57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07" Dec 03 14:21:20 crc kubenswrapper[4677]: I1203 14:21:20.030275 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07"} err="failed to get container status \"57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07\": rpc error: code = NotFound desc = could not find container \"57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07\": container with ID starting with 57f68cb6fa7914a2587dfd59a3dab2bcb33d609a22ca640479a59e6757676b07 not found: ID does not exist" Dec 03 14:21:20 crc kubenswrapper[4677]: I1203 14:21:20.030305 4677 scope.go:117] "RemoveContainer" containerID="e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966" Dec 03 14:21:20 crc kubenswrapper[4677]: E1203 14:21:20.030797 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966\": container with ID starting with e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966 not found: ID does not exist" containerID="e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966" Dec 03 14:21:20 crc kubenswrapper[4677]: I1203 14:21:20.030831 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966"} err="failed to get container status \"e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966\": rpc error: code = NotFound desc = could not find container \"e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966\": container with ID starting with e2dd99e16a630980529a6f5663b9a5f52cf6debbdf4995748ed1f19039258966 not found: ID does not exist" Dec 03 14:21:21 crc kubenswrapper[4677]: I1203 14:21:21.992223 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" path="/var/lib/kubelet/pods/b57e4890-c847-427c-a4cc-b11bab6891c2/volumes" Dec 03 14:21:38 crc kubenswrapper[4677]: I1203 14:21:38.437685 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:21:38 crc kubenswrapper[4677]: I1203 14:21:38.438322 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:21:44 crc kubenswrapper[4677]: I1203 14:21:44.185346 4677 generic.go:334] "Generic (PLEG): container finished" podID="21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737" containerID="1bb3cab2f91e90c85f97b83321a8b589885c42b1c6504c6c9e839f91a665a3ee" exitCode=0 Dec 03 14:21:44 crc kubenswrapper[4677]: I1203 14:21:44.185448 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" event={"ID":"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737","Type":"ContainerDied","Data":"1bb3cab2f91e90c85f97b83321a8b589885c42b1c6504c6c9e839f91a665a3ee"} Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.642368 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.672232 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-ssh-key\") pod \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.672316 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmz7m\" (UniqueName: \"kubernetes.io/projected/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-kube-api-access-hmz7m\") pod \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.672457 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-inventory\") pod \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\" (UID: \"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737\") " Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.686457 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-kube-api-access-hmz7m" (OuterVolumeSpecName: "kube-api-access-hmz7m") pod "21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737" (UID: "21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737"). InnerVolumeSpecName "kube-api-access-hmz7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.702187 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737" (UID: "21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.707893 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-inventory" (OuterVolumeSpecName: "inventory") pod "21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737" (UID: "21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.774907 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmz7m\" (UniqueName: \"kubernetes.io/projected/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-kube-api-access-hmz7m\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.774941 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:45 crc kubenswrapper[4677]: I1203 14:21:45.774963 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.218444 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.219895 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-25g8q" event={"ID":"21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737","Type":"ContainerDied","Data":"495e72f71d1eefb84cb827b61cee7409142a9414f416e1ce7f0dcef7385389d2"} Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.219941 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="495e72f71d1eefb84cb827b61cee7409142a9414f416e1ce7f0dcef7385389d2" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341120 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv"] Dec 03 14:21:46 crc kubenswrapper[4677]: E1203 14:21:46.341589 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341620 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 14:21:46 crc kubenswrapper[4677]: E1203 14:21:46.341664 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="registry-server" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341674 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="registry-server" Dec 03 14:21:46 crc kubenswrapper[4677]: E1203 14:21:46.341695 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="extract-content" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341704 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="extract-content" Dec 03 14:21:46 crc kubenswrapper[4677]: E1203 14:21:46.341724 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="registry-server" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341731 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="registry-server" Dec 03 14:21:46 crc kubenswrapper[4677]: E1203 14:21:46.341747 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="extract-utilities" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341756 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="extract-utilities" Dec 03 14:21:46 crc kubenswrapper[4677]: E1203 14:21:46.341769 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="extract-utilities" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341776 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="extract-utilities" Dec 03 14:21:46 crc kubenswrapper[4677]: E1203 14:21:46.341801 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="extract-content" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.341809 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="extract-content" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.342103 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="13baedf7-2496-48d4-b165-1efddc4e20a0" containerName="registry-server" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.342156 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b57e4890-c847-427c-a4cc-b11bab6891c2" containerName="registry-server" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.342186 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.343127 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.348514 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.348692 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.348815 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.348933 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.355015 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv"] Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.390613 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.390841 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-247k5\" (UniqueName: \"kubernetes.io/projected/85524a2c-13a3-494d-8528-9a92c581153b-kube-api-access-247k5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.390883 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.491550 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-247k5\" (UniqueName: \"kubernetes.io/projected/85524a2c-13a3-494d-8528-9a92c581153b-kube-api-access-247k5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.491594 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.491639 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.497077 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.497095 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.507705 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-247k5\" (UniqueName: \"kubernetes.io/projected/85524a2c-13a3-494d-8528-9a92c581153b-kube-api-access-247k5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-r4brv\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:46 crc kubenswrapper[4677]: I1203 14:21:46.667361 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:21:47 crc kubenswrapper[4677]: I1203 14:21:47.162819 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv"] Dec 03 14:21:47 crc kubenswrapper[4677]: I1203 14:21:47.230705 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" event={"ID":"85524a2c-13a3-494d-8528-9a92c581153b","Type":"ContainerStarted","Data":"843dcb2aa0398e124157e0611d1703d7c4fe12e061722570fbf6af16a9272602"} Dec 03 14:21:48 crc kubenswrapper[4677]: I1203 14:21:48.240705 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" event={"ID":"85524a2c-13a3-494d-8528-9a92c581153b","Type":"ContainerStarted","Data":"3186623dbc32b16a93fe1ee3c4cb1de0d050a20d05c3a4c5d6829516806ca6ec"} Dec 03 14:21:48 crc kubenswrapper[4677]: I1203 14:21:48.262298 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" podStartSLOduration=1.8651481840000002 podStartE2EDuration="2.26227861s" podCreationTimestamp="2025-12-03 14:21:46 +0000 UTC" firstStartedPulling="2025-12-03 14:21:47.174003817 +0000 UTC m=+2097.920336272" lastFinishedPulling="2025-12-03 14:21:47.571134243 +0000 UTC m=+2098.317466698" observedRunningTime="2025-12-03 14:21:48.256533354 +0000 UTC m=+2099.002865819" watchObservedRunningTime="2025-12-03 14:21:48.26227861 +0000 UTC m=+2099.008611065" Dec 03 14:21:50 crc kubenswrapper[4677]: I1203 14:21:50.046856 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-nm2zs"] Dec 03 14:21:50 crc kubenswrapper[4677]: I1203 14:21:50.076617 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-csbsv"] Dec 03 14:21:50 crc kubenswrapper[4677]: I1203 14:21:50.088276 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-csbsv"] Dec 03 14:21:50 crc kubenswrapper[4677]: I1203 14:21:50.097265 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-nm2zs"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.041142 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-vgwg2"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.050443 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-27be-account-create-update-v69nt"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.060733 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b4d9-account-create-update-dv852"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.068784 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9792-account-create-update-8k85x"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.076109 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-vgwg2"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.083478 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b4d9-account-create-update-dv852"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.092468 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9792-account-create-update-8k85x"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.099794 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-27be-account-create-update-v69nt"] Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.989237 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11d47538-6f9f-47ec-8d63-ddc3abb6f7a8" path="/var/lib/kubelet/pods/11d47538-6f9f-47ec-8d63-ddc3abb6f7a8/volumes" Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.990414 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23c51891-2a54-4a2b-8af7-9e18c571d3ea" path="/var/lib/kubelet/pods/23c51891-2a54-4a2b-8af7-9e18c571d3ea/volumes" Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.991038 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48e57ed4-f60c-4bd1-a9dc-68a298bd76a5" path="/var/lib/kubelet/pods/48e57ed4-f60c-4bd1-a9dc-68a298bd76a5/volumes" Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.991885 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c23c782-4305-4cb7-a855-2f6dd16c45e2" path="/var/lib/kubelet/pods/5c23c782-4305-4cb7-a855-2f6dd16c45e2/volumes" Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.993302 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d232644-1f01-4ed4-9ca5-d4259a0d1c99" path="/var/lib/kubelet/pods/9d232644-1f01-4ed4-9ca5-d4259a0d1c99/volumes" Dec 03 14:21:51 crc kubenswrapper[4677]: I1203 14:21:51.993839 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3baef30-7171-433b-95c7-1810d7bb41a1" path="/var/lib/kubelet/pods/e3baef30-7171-433b-95c7-1810d7bb41a1/volumes" Dec 03 14:22:06 crc kubenswrapper[4677]: I1203 14:22:06.158412 4677 scope.go:117] "RemoveContainer" containerID="06ffbe4a2a5ac7704b579d41f6e9b7ab32d7c405ec971c358f3c2dbba2c4438b" Dec 03 14:22:06 crc kubenswrapper[4677]: I1203 14:22:06.196091 4677 scope.go:117] "RemoveContainer" containerID="21f066640d6b3a2cfffdaabc936b8a54af82b091cc1ed1cc841f91b39eff42f8" Dec 03 14:22:06 crc kubenswrapper[4677]: I1203 14:22:06.242805 4677 scope.go:117] "RemoveContainer" containerID="44b07641e6b0ad63aedc7f27be4b00c065566e9e80979d0646d57822cb1ab71c" Dec 03 14:22:06 crc kubenswrapper[4677]: I1203 14:22:06.298411 4677 scope.go:117] "RemoveContainer" containerID="5304c636ef19ee93e710fd284c2246113003e33b289706562035a1d5ebae6bbc" Dec 03 14:22:06 crc kubenswrapper[4677]: I1203 14:22:06.342858 4677 scope.go:117] "RemoveContainer" containerID="fd003a296f156a154b9515e27c712505c2a2f1e48653f486ff3ea622268e3325" Dec 03 14:22:06 crc kubenswrapper[4677]: I1203 14:22:06.390526 4677 scope.go:117] "RemoveContainer" containerID="79e2f298d3a112fd3e9246eb9d90433b251279eea18af696131d0c46d7baeeeb" Dec 03 14:22:08 crc kubenswrapper[4677]: I1203 14:22:08.436937 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:22:08 crc kubenswrapper[4677]: I1203 14:22:08.437382 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:22:08 crc kubenswrapper[4677]: I1203 14:22:08.437427 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:22:08 crc kubenswrapper[4677]: I1203 14:22:08.438172 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc95799bd3832da5bee136d530caa6131c382dcae6b9ac4b24eb38e3ea0e1e21"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:22:08 crc kubenswrapper[4677]: I1203 14:22:08.438229 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://dc95799bd3832da5bee136d530caa6131c382dcae6b9ac4b24eb38e3ea0e1e21" gracePeriod=600 Dec 03 14:22:09 crc kubenswrapper[4677]: I1203 14:22:09.441797 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="dc95799bd3832da5bee136d530caa6131c382dcae6b9ac4b24eb38e3ea0e1e21" exitCode=0 Dec 03 14:22:09 crc kubenswrapper[4677]: I1203 14:22:09.441886 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"dc95799bd3832da5bee136d530caa6131c382dcae6b9ac4b24eb38e3ea0e1e21"} Dec 03 14:22:09 crc kubenswrapper[4677]: I1203 14:22:09.442157 4677 scope.go:117] "RemoveContainer" containerID="48e087a65c3344e73e6c7d0e5a446930cf6e334fca95b34fc98a8f92f0d653f3" Dec 03 14:22:10 crc kubenswrapper[4677]: I1203 14:22:10.453901 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e"} Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.629575 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-npfdc"] Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.634740 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.646283 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-npfdc"] Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.672699 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-catalog-content\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.673009 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvg58\" (UniqueName: \"kubernetes.io/projected/9e70821a-d352-4647-adf1-7b87efe889da-kube-api-access-zvg58\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.673248 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-utilities\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.774466 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-utilities\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.774570 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-catalog-content\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.774619 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvg58\" (UniqueName: \"kubernetes.io/projected/9e70821a-d352-4647-adf1-7b87efe889da-kube-api-access-zvg58\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.775186 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-utilities\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.775201 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-catalog-content\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.796875 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvg58\" (UniqueName: \"kubernetes.io/projected/9e70821a-d352-4647-adf1-7b87efe889da-kube-api-access-zvg58\") pod \"redhat-operators-npfdc\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:21 crc kubenswrapper[4677]: I1203 14:22:21.962028 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:22 crc kubenswrapper[4677]: I1203 14:22:22.469771 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-npfdc"] Dec 03 14:22:22 crc kubenswrapper[4677]: W1203 14:22:22.488924 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e70821a_d352_4647_adf1_7b87efe889da.slice/crio-3d56488cf6d1963657ad55efd79a7a045fc2dd59fab93d45e988d707fd85cc5e WatchSource:0}: Error finding container 3d56488cf6d1963657ad55efd79a7a045fc2dd59fab93d45e988d707fd85cc5e: Status 404 returned error can't find the container with id 3d56488cf6d1963657ad55efd79a7a045fc2dd59fab93d45e988d707fd85cc5e Dec 03 14:22:22 crc kubenswrapper[4677]: I1203 14:22:22.556417 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npfdc" event={"ID":"9e70821a-d352-4647-adf1-7b87efe889da","Type":"ContainerStarted","Data":"3d56488cf6d1963657ad55efd79a7a045fc2dd59fab93d45e988d707fd85cc5e"} Dec 03 14:22:23 crc kubenswrapper[4677]: I1203 14:22:23.567130 4677 generic.go:334] "Generic (PLEG): container finished" podID="9e70821a-d352-4647-adf1-7b87efe889da" containerID="0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99" exitCode=0 Dec 03 14:22:23 crc kubenswrapper[4677]: I1203 14:22:23.567178 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npfdc" event={"ID":"9e70821a-d352-4647-adf1-7b87efe889da","Type":"ContainerDied","Data":"0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99"} Dec 03 14:22:24 crc kubenswrapper[4677]: I1203 14:22:24.583252 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npfdc" event={"ID":"9e70821a-d352-4647-adf1-7b87efe889da","Type":"ContainerStarted","Data":"54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92"} Dec 03 14:22:25 crc kubenswrapper[4677]: I1203 14:22:25.053151 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkxw5"] Dec 03 14:22:25 crc kubenswrapper[4677]: I1203 14:22:25.061801 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkxw5"] Dec 03 14:22:26 crc kubenswrapper[4677]: I1203 14:22:26.021693 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b349d0-d7aa-45cb-aab4-d50379b27c4d" path="/var/lib/kubelet/pods/b9b349d0-d7aa-45cb-aab4-d50379b27c4d/volumes" Dec 03 14:22:28 crc kubenswrapper[4677]: I1203 14:22:28.619147 4677 generic.go:334] "Generic (PLEG): container finished" podID="9e70821a-d352-4647-adf1-7b87efe889da" containerID="54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92" exitCode=0 Dec 03 14:22:28 crc kubenswrapper[4677]: I1203 14:22:28.619186 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npfdc" event={"ID":"9e70821a-d352-4647-adf1-7b87efe889da","Type":"ContainerDied","Data":"54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92"} Dec 03 14:22:29 crc kubenswrapper[4677]: I1203 14:22:29.645163 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npfdc" event={"ID":"9e70821a-d352-4647-adf1-7b87efe889da","Type":"ContainerStarted","Data":"4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202"} Dec 03 14:22:29 crc kubenswrapper[4677]: I1203 14:22:29.674532 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-npfdc" podStartSLOduration=3.044072739 podStartE2EDuration="8.674513048s" podCreationTimestamp="2025-12-03 14:22:21 +0000 UTC" firstStartedPulling="2025-12-03 14:22:23.568845823 +0000 UTC m=+2134.315178278" lastFinishedPulling="2025-12-03 14:22:29.199286112 +0000 UTC m=+2139.945618587" observedRunningTime="2025-12-03 14:22:29.67045757 +0000 UTC m=+2140.416790035" watchObservedRunningTime="2025-12-03 14:22:29.674513048 +0000 UTC m=+2140.420845493" Dec 03 14:22:31 crc kubenswrapper[4677]: I1203 14:22:31.962665 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:31 crc kubenswrapper[4677]: I1203 14:22:31.963026 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:33 crc kubenswrapper[4677]: I1203 14:22:33.014908 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-npfdc" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="registry-server" probeResult="failure" output=< Dec 03 14:22:33 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 14:22:33 crc kubenswrapper[4677]: > Dec 03 14:22:42 crc kubenswrapper[4677]: I1203 14:22:42.020572 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:42 crc kubenswrapper[4677]: I1203 14:22:42.080449 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:42 crc kubenswrapper[4677]: I1203 14:22:42.253101 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-npfdc"] Dec 03 14:22:43 crc kubenswrapper[4677]: I1203 14:22:43.771785 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-npfdc" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="registry-server" containerID="cri-o://4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202" gracePeriod=2 Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.287414 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.409262 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-catalog-content\") pod \"9e70821a-d352-4647-adf1-7b87efe889da\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.409311 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvg58\" (UniqueName: \"kubernetes.io/projected/9e70821a-d352-4647-adf1-7b87efe889da-kube-api-access-zvg58\") pod \"9e70821a-d352-4647-adf1-7b87efe889da\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.409483 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-utilities\") pod \"9e70821a-d352-4647-adf1-7b87efe889da\" (UID: \"9e70821a-d352-4647-adf1-7b87efe889da\") " Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.410830 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-utilities" (OuterVolumeSpecName: "utilities") pod "9e70821a-d352-4647-adf1-7b87efe889da" (UID: "9e70821a-d352-4647-adf1-7b87efe889da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.418786 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e70821a-d352-4647-adf1-7b87efe889da-kube-api-access-zvg58" (OuterVolumeSpecName: "kube-api-access-zvg58") pod "9e70821a-d352-4647-adf1-7b87efe889da" (UID: "9e70821a-d352-4647-adf1-7b87efe889da"). InnerVolumeSpecName "kube-api-access-zvg58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.511683 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvg58\" (UniqueName: \"kubernetes.io/projected/9e70821a-d352-4647-adf1-7b87efe889da-kube-api-access-zvg58\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.511715 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.528759 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e70821a-d352-4647-adf1-7b87efe889da" (UID: "9e70821a-d352-4647-adf1-7b87efe889da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.614131 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e70821a-d352-4647-adf1-7b87efe889da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.781582 4677 generic.go:334] "Generic (PLEG): container finished" podID="9e70821a-d352-4647-adf1-7b87efe889da" containerID="4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202" exitCode=0 Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.781624 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npfdc" event={"ID":"9e70821a-d352-4647-adf1-7b87efe889da","Type":"ContainerDied","Data":"4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202"} Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.781659 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npfdc" event={"ID":"9e70821a-d352-4647-adf1-7b87efe889da","Type":"ContainerDied","Data":"3d56488cf6d1963657ad55efd79a7a045fc2dd59fab93d45e988d707fd85cc5e"} Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.781680 4677 scope.go:117] "RemoveContainer" containerID="4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.783500 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npfdc" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.802642 4677 scope.go:117] "RemoveContainer" containerID="54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.848019 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-npfdc"] Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.860682 4677 scope.go:117] "RemoveContainer" containerID="0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.863016 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-npfdc"] Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.900720 4677 scope.go:117] "RemoveContainer" containerID="4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202" Dec 03 14:22:44 crc kubenswrapper[4677]: E1203 14:22:44.902021 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202\": container with ID starting with 4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202 not found: ID does not exist" containerID="4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.902050 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202"} err="failed to get container status \"4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202\": rpc error: code = NotFound desc = could not find container \"4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202\": container with ID starting with 4c8fc54b3c14fa6d393fbd0d02ba7610f5801878fe35d4184dce1c8bb66e3202 not found: ID does not exist" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.902078 4677 scope.go:117] "RemoveContainer" containerID="54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92" Dec 03 14:22:44 crc kubenswrapper[4677]: E1203 14:22:44.902470 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92\": container with ID starting with 54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92 not found: ID does not exist" containerID="54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.902516 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92"} err="failed to get container status \"54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92\": rpc error: code = NotFound desc = could not find container \"54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92\": container with ID starting with 54be7492df68e144cd32ec5de7d61b2537319180c717d80337c282407e5aaa92 not found: ID does not exist" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.902546 4677 scope.go:117] "RemoveContainer" containerID="0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99" Dec 03 14:22:44 crc kubenswrapper[4677]: E1203 14:22:44.902847 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99\": container with ID starting with 0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99 not found: ID does not exist" containerID="0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99" Dec 03 14:22:44 crc kubenswrapper[4677]: I1203 14:22:44.902877 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99"} err="failed to get container status \"0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99\": rpc error: code = NotFound desc = could not find container \"0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99\": container with ID starting with 0863d5166e0828798979b7c26393b95f9ac10ddf7236d7efb0a3f673ccbf6a99 not found: ID does not exist" Dec 03 14:22:45 crc kubenswrapper[4677]: I1203 14:22:45.987701 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e70821a-d352-4647-adf1-7b87efe889da" path="/var/lib/kubelet/pods/9e70821a-d352-4647-adf1-7b87efe889da/volumes" Dec 03 14:22:48 crc kubenswrapper[4677]: I1203 14:22:48.032145 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-kjshq"] Dec 03 14:22:48 crc kubenswrapper[4677]: I1203 14:22:48.041829 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-kjshq"] Dec 03 14:22:50 crc kubenswrapper[4677]: I1203 14:22:50.040729 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fed8b246-9742-4b78-a630-b820a90ec8c8" path="/var/lib/kubelet/pods/fed8b246-9742-4b78-a630-b820a90ec8c8/volumes" Dec 03 14:22:53 crc kubenswrapper[4677]: I1203 14:22:53.049090 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9vwd9"] Dec 03 14:22:53 crc kubenswrapper[4677]: I1203 14:22:53.061426 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9vwd9"] Dec 03 14:22:53 crc kubenswrapper[4677]: I1203 14:22:53.987884 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42fd76ce-441e-4c60-9220-89d31123f933" path="/var/lib/kubelet/pods/42fd76ce-441e-4c60-9220-89d31123f933/volumes" Dec 03 14:23:06 crc kubenswrapper[4677]: I1203 14:23:06.590438 4677 scope.go:117] "RemoveContainer" containerID="580f81c7408191a5eb21928675151e5550399ebe1ed8cbbe07a3f0aa9034385c" Dec 03 14:23:06 crc kubenswrapper[4677]: I1203 14:23:06.647767 4677 scope.go:117] "RemoveContainer" containerID="047d9dc6493b6bc1a221b6c2bdad811a81529f68519c088a20b717a7fd6bbd15" Dec 03 14:23:06 crc kubenswrapper[4677]: I1203 14:23:06.705997 4677 scope.go:117] "RemoveContainer" containerID="2f225efbd9d3e9109b5f8d382b1e486ab86ba270b109f95891c04c3490e486f4" Dec 03 14:23:34 crc kubenswrapper[4677]: I1203 14:23:34.048649 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-wxzts"] Dec 03 14:23:34 crc kubenswrapper[4677]: I1203 14:23:34.058619 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-wxzts"] Dec 03 14:23:35 crc kubenswrapper[4677]: I1203 14:23:35.986673 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf" path="/var/lib/kubelet/pods/ddc0cadd-ec0b-4fbf-841b-cb88c29aeacf/volumes" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.186459 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nbc7m"] Dec 03 14:23:53 crc kubenswrapper[4677]: E1203 14:23:53.187564 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="registry-server" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.187586 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="registry-server" Dec 03 14:23:53 crc kubenswrapper[4677]: E1203 14:23:53.187626 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="extract-utilities" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.187635 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="extract-utilities" Dec 03 14:23:53 crc kubenswrapper[4677]: E1203 14:23:53.187651 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="extract-content" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.187659 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="extract-content" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.187972 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e70821a-d352-4647-adf1-7b87efe889da" containerName="registry-server" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.189986 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.215821 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nbc7m"] Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.300767 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-utilities\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.300844 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2942\" (UniqueName: \"kubernetes.io/projected/701f2ae9-c397-4e5e-a340-b2abc769990c-kube-api-access-z2942\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.300871 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-catalog-content\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.402805 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-utilities\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.402917 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2942\" (UniqueName: \"kubernetes.io/projected/701f2ae9-c397-4e5e-a340-b2abc769990c-kube-api-access-z2942\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.402968 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-catalog-content\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.403697 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-utilities\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.403809 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-catalog-content\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.433046 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2942\" (UniqueName: \"kubernetes.io/projected/701f2ae9-c397-4e5e-a340-b2abc769990c-kube-api-access-z2942\") pod \"certified-operators-nbc7m\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:53 crc kubenswrapper[4677]: I1203 14:23:53.525491 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:23:54 crc kubenswrapper[4677]: I1203 14:23:54.062038 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nbc7m"] Dec 03 14:23:54 crc kubenswrapper[4677]: I1203 14:23:54.411752 4677 generic.go:334] "Generic (PLEG): container finished" podID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerID="83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f" exitCode=0 Dec 03 14:23:54 crc kubenswrapper[4677]: I1203 14:23:54.411815 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbc7m" event={"ID":"701f2ae9-c397-4e5e-a340-b2abc769990c","Type":"ContainerDied","Data":"83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f"} Dec 03 14:23:54 crc kubenswrapper[4677]: I1203 14:23:54.412192 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbc7m" event={"ID":"701f2ae9-c397-4e5e-a340-b2abc769990c","Type":"ContainerStarted","Data":"dce08f05fb8d6fd7cb0c5a9f23241d99965bf53b305ed5084d8fcc503bfc24e0"} Dec 03 14:23:54 crc kubenswrapper[4677]: I1203 14:23:54.414518 4677 generic.go:334] "Generic (PLEG): container finished" podID="85524a2c-13a3-494d-8528-9a92c581153b" containerID="3186623dbc32b16a93fe1ee3c4cb1de0d050a20d05c3a4c5d6829516806ca6ec" exitCode=0 Dec 03 14:23:54 crc kubenswrapper[4677]: I1203 14:23:54.414620 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" event={"ID":"85524a2c-13a3-494d-8528-9a92c581153b","Type":"ContainerDied","Data":"3186623dbc32b16a93fe1ee3c4cb1de0d050a20d05c3a4c5d6829516806ca6ec"} Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.425361 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbc7m" event={"ID":"701f2ae9-c397-4e5e-a340-b2abc769990c","Type":"ContainerStarted","Data":"24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8"} Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.903888 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.956054 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-ssh-key\") pod \"85524a2c-13a3-494d-8528-9a92c581153b\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.956452 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-247k5\" (UniqueName: \"kubernetes.io/projected/85524a2c-13a3-494d-8528-9a92c581153b-kube-api-access-247k5\") pod \"85524a2c-13a3-494d-8528-9a92c581153b\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.956777 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-inventory\") pod \"85524a2c-13a3-494d-8528-9a92c581153b\" (UID: \"85524a2c-13a3-494d-8528-9a92c581153b\") " Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.962234 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85524a2c-13a3-494d-8528-9a92c581153b-kube-api-access-247k5" (OuterVolumeSpecName: "kube-api-access-247k5") pod "85524a2c-13a3-494d-8528-9a92c581153b" (UID: "85524a2c-13a3-494d-8528-9a92c581153b"). InnerVolumeSpecName "kube-api-access-247k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.988059 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "85524a2c-13a3-494d-8528-9a92c581153b" (UID: "85524a2c-13a3-494d-8528-9a92c581153b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:23:55 crc kubenswrapper[4677]: I1203 14:23:55.994199 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-inventory" (OuterVolumeSpecName: "inventory") pod "85524a2c-13a3-494d-8528-9a92c581153b" (UID: "85524a2c-13a3-494d-8528-9a92c581153b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.059204 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.059244 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85524a2c-13a3-494d-8528-9a92c581153b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.059258 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-247k5\" (UniqueName: \"kubernetes.io/projected/85524a2c-13a3-494d-8528-9a92c581153b-kube-api-access-247k5\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.434320 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" event={"ID":"85524a2c-13a3-494d-8528-9a92c581153b","Type":"ContainerDied","Data":"843dcb2aa0398e124157e0611d1703d7c4fe12e061722570fbf6af16a9272602"} Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.434374 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="843dcb2aa0398e124157e0611d1703d7c4fe12e061722570fbf6af16a9272602" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.434337 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-r4brv" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.438141 4677 generic.go:334] "Generic (PLEG): container finished" podID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerID="24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8" exitCode=0 Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.438175 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbc7m" event={"ID":"701f2ae9-c397-4e5e-a340-b2abc769990c","Type":"ContainerDied","Data":"24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8"} Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.549071 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr"] Dec 03 14:23:56 crc kubenswrapper[4677]: E1203 14:23:56.549625 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85524a2c-13a3-494d-8528-9a92c581153b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.549659 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="85524a2c-13a3-494d-8528-9a92c581153b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.549916 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="85524a2c-13a3-494d-8528-9a92c581153b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.550857 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.554144 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.554905 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.555173 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.555231 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.568129 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr"] Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.670834 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.671109 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbvfv\" (UniqueName: \"kubernetes.io/projected/aa53aac3-038e-4728-bbec-afce4e5c7509-kube-api-access-rbvfv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.671468 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.773510 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.773646 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.773770 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbvfv\" (UniqueName: \"kubernetes.io/projected/aa53aac3-038e-4728-bbec-afce4e5c7509-kube-api-access-rbvfv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.777218 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.781537 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.789780 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbvfv\" (UniqueName: \"kubernetes.io/projected/aa53aac3-038e-4728-bbec-afce4e5c7509-kube-api-access-rbvfv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:56 crc kubenswrapper[4677]: I1203 14:23:56.868581 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:23:57 crc kubenswrapper[4677]: I1203 14:23:57.414167 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr"] Dec 03 14:23:57 crc kubenswrapper[4677]: I1203 14:23:57.447619 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbc7m" event={"ID":"701f2ae9-c397-4e5e-a340-b2abc769990c","Type":"ContainerStarted","Data":"c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8"} Dec 03 14:23:57 crc kubenswrapper[4677]: I1203 14:23:57.449601 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" event={"ID":"aa53aac3-038e-4728-bbec-afce4e5c7509","Type":"ContainerStarted","Data":"c893507d1e80e36879dd85246bcf3ac8e436fd4ce423f5a1ef079405361c28a6"} Dec 03 14:23:57 crc kubenswrapper[4677]: I1203 14:23:57.469960 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nbc7m" podStartSLOduration=2.025401482 podStartE2EDuration="4.469932963s" podCreationTimestamp="2025-12-03 14:23:53 +0000 UTC" firstStartedPulling="2025-12-03 14:23:54.416206826 +0000 UTC m=+2225.162539291" lastFinishedPulling="2025-12-03 14:23:56.860738317 +0000 UTC m=+2227.607070772" observedRunningTime="2025-12-03 14:23:57.468374441 +0000 UTC m=+2228.214706906" watchObservedRunningTime="2025-12-03 14:23:57.469932963 +0000 UTC m=+2228.216265418" Dec 03 14:23:58 crc kubenswrapper[4677]: I1203 14:23:58.462155 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" event={"ID":"aa53aac3-038e-4728-bbec-afce4e5c7509","Type":"ContainerStarted","Data":"1ccc73a6b8f7cbd4c7b80476e68e35d5d32039afcf15046ac00928566e5b70e1"} Dec 03 14:23:58 crc kubenswrapper[4677]: I1203 14:23:58.478349 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" podStartSLOduration=2.040293172 podStartE2EDuration="2.478325091s" podCreationTimestamp="2025-12-03 14:23:56 +0000 UTC" firstStartedPulling="2025-12-03 14:23:57.404509754 +0000 UTC m=+2228.150842209" lastFinishedPulling="2025-12-03 14:23:57.842541673 +0000 UTC m=+2228.588874128" observedRunningTime="2025-12-03 14:23:58.477894019 +0000 UTC m=+2229.224226474" watchObservedRunningTime="2025-12-03 14:23:58.478325091 +0000 UTC m=+2229.224657576" Dec 03 14:24:03 crc kubenswrapper[4677]: I1203 14:24:03.511987 4677 generic.go:334] "Generic (PLEG): container finished" podID="aa53aac3-038e-4728-bbec-afce4e5c7509" containerID="1ccc73a6b8f7cbd4c7b80476e68e35d5d32039afcf15046ac00928566e5b70e1" exitCode=0 Dec 03 14:24:03 crc kubenswrapper[4677]: I1203 14:24:03.512354 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" event={"ID":"aa53aac3-038e-4728-bbec-afce4e5c7509","Type":"ContainerDied","Data":"1ccc73a6b8f7cbd4c7b80476e68e35d5d32039afcf15046ac00928566e5b70e1"} Dec 03 14:24:03 crc kubenswrapper[4677]: I1203 14:24:03.527985 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:24:03 crc kubenswrapper[4677]: I1203 14:24:03.528395 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:24:03 crc kubenswrapper[4677]: I1203 14:24:03.585355 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:24:04 crc kubenswrapper[4677]: I1203 14:24:04.572404 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:24:04 crc kubenswrapper[4677]: I1203 14:24:04.924671 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.033814 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-ssh-key\") pod \"aa53aac3-038e-4728-bbec-afce4e5c7509\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.034015 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbvfv\" (UniqueName: \"kubernetes.io/projected/aa53aac3-038e-4728-bbec-afce4e5c7509-kube-api-access-rbvfv\") pod \"aa53aac3-038e-4728-bbec-afce4e5c7509\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.034084 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-inventory\") pod \"aa53aac3-038e-4728-bbec-afce4e5c7509\" (UID: \"aa53aac3-038e-4728-bbec-afce4e5c7509\") " Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.043259 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa53aac3-038e-4728-bbec-afce4e5c7509-kube-api-access-rbvfv" (OuterVolumeSpecName: "kube-api-access-rbvfv") pod "aa53aac3-038e-4728-bbec-afce4e5c7509" (UID: "aa53aac3-038e-4728-bbec-afce4e5c7509"). InnerVolumeSpecName "kube-api-access-rbvfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.061837 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa53aac3-038e-4728-bbec-afce4e5c7509" (UID: "aa53aac3-038e-4728-bbec-afce4e5c7509"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.102937 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-inventory" (OuterVolumeSpecName: "inventory") pod "aa53aac3-038e-4728-bbec-afce4e5c7509" (UID: "aa53aac3-038e-4728-bbec-afce4e5c7509"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.137253 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.137319 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbvfv\" (UniqueName: \"kubernetes.io/projected/aa53aac3-038e-4728-bbec-afce4e5c7509-kube-api-access-rbvfv\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.137376 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa53aac3-038e-4728-bbec-afce4e5c7509-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.534085 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" event={"ID":"aa53aac3-038e-4728-bbec-afce4e5c7509","Type":"ContainerDied","Data":"c893507d1e80e36879dd85246bcf3ac8e436fd4ce423f5a1ef079405361c28a6"} Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.534117 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.534129 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c893507d1e80e36879dd85246bcf3ac8e436fd4ce423f5a1ef079405361c28a6" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.611034 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k"] Dec 03 14:24:05 crc kubenswrapper[4677]: E1203 14:24:05.611451 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa53aac3-038e-4728-bbec-afce4e5c7509" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.611467 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa53aac3-038e-4728-bbec-afce4e5c7509" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.611729 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa53aac3-038e-4728-bbec-afce4e5c7509" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.612581 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.616748 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.616980 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.617188 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.617333 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.624416 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k"] Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.650597 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9xmp\" (UniqueName: \"kubernetes.io/projected/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-kube-api-access-j9xmp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.650704 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.650751 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.753435 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9xmp\" (UniqueName: \"kubernetes.io/projected/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-kube-api-access-j9xmp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.753577 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.753625 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.759584 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.760624 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.775851 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9xmp\" (UniqueName: \"kubernetes.io/projected/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-kube-api-access-j9xmp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5q8k\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:05 crc kubenswrapper[4677]: I1203 14:24:05.941625 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:06 crc kubenswrapper[4677]: I1203 14:24:06.510404 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k"] Dec 03 14:24:06 crc kubenswrapper[4677]: W1203 14:24:06.514830 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c6dfb73_f46a_4233_a1bf_a22c6e9e923e.slice/crio-a79bf165a3c661c212bd2c65182096b18af1e5078f304ffe6faefb592f2369ca WatchSource:0}: Error finding container a79bf165a3c661c212bd2c65182096b18af1e5078f304ffe6faefb592f2369ca: Status 404 returned error can't find the container with id a79bf165a3c661c212bd2c65182096b18af1e5078f304ffe6faefb592f2369ca Dec 03 14:24:06 crc kubenswrapper[4677]: I1203 14:24:06.546928 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" event={"ID":"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e","Type":"ContainerStarted","Data":"a79bf165a3c661c212bd2c65182096b18af1e5078f304ffe6faefb592f2369ca"} Dec 03 14:24:06 crc kubenswrapper[4677]: I1203 14:24:06.862777 4677 scope.go:117] "RemoveContainer" containerID="85df570ee32a8fd9b37707d8d51e419d2f0db38d23b67a8e9dc513992a8ea6cc" Dec 03 14:24:07 crc kubenswrapper[4677]: I1203 14:24:07.172846 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nbc7m"] Dec 03 14:24:07 crc kubenswrapper[4677]: I1203 14:24:07.561036 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" event={"ID":"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e","Type":"ContainerStarted","Data":"44e04b3fe97061d4910405e3133dd02b8d1f7f6561b47082475b73c1ba989f8d"} Dec 03 14:24:07 crc kubenswrapper[4677]: I1203 14:24:07.561335 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nbc7m" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="registry-server" containerID="cri-o://c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8" gracePeriod=2 Dec 03 14:24:07 crc kubenswrapper[4677]: I1203 14:24:07.578351 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" podStartSLOduration=2.037659871 podStartE2EDuration="2.578324553s" podCreationTimestamp="2025-12-03 14:24:05 +0000 UTC" firstStartedPulling="2025-12-03 14:24:06.516938228 +0000 UTC m=+2237.263270683" lastFinishedPulling="2025-12-03 14:24:07.05760292 +0000 UTC m=+2237.803935365" observedRunningTime="2025-12-03 14:24:07.577056678 +0000 UTC m=+2238.323389133" watchObservedRunningTime="2025-12-03 14:24:07.578324553 +0000 UTC m=+2238.324657008" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.038084 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.118973 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-utilities\") pod \"701f2ae9-c397-4e5e-a340-b2abc769990c\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.119185 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2942\" (UniqueName: \"kubernetes.io/projected/701f2ae9-c397-4e5e-a340-b2abc769990c-kube-api-access-z2942\") pod \"701f2ae9-c397-4e5e-a340-b2abc769990c\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.119218 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-catalog-content\") pod \"701f2ae9-c397-4e5e-a340-b2abc769990c\" (UID: \"701f2ae9-c397-4e5e-a340-b2abc769990c\") " Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.120072 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-utilities" (OuterVolumeSpecName: "utilities") pod "701f2ae9-c397-4e5e-a340-b2abc769990c" (UID: "701f2ae9-c397-4e5e-a340-b2abc769990c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.128368 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/701f2ae9-c397-4e5e-a340-b2abc769990c-kube-api-access-z2942" (OuterVolumeSpecName: "kube-api-access-z2942") pod "701f2ae9-c397-4e5e-a340-b2abc769990c" (UID: "701f2ae9-c397-4e5e-a340-b2abc769990c"). InnerVolumeSpecName "kube-api-access-z2942". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.184396 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "701f2ae9-c397-4e5e-a340-b2abc769990c" (UID: "701f2ae9-c397-4e5e-a340-b2abc769990c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.221404 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.221436 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2942\" (UniqueName: \"kubernetes.io/projected/701f2ae9-c397-4e5e-a340-b2abc769990c-kube-api-access-z2942\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.221446 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/701f2ae9-c397-4e5e-a340-b2abc769990c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.572461 4677 generic.go:334] "Generic (PLEG): container finished" podID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerID="c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8" exitCode=0 Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.572521 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbc7m" event={"ID":"701f2ae9-c397-4e5e-a340-b2abc769990c","Type":"ContainerDied","Data":"c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8"} Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.572805 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbc7m" event={"ID":"701f2ae9-c397-4e5e-a340-b2abc769990c","Type":"ContainerDied","Data":"dce08f05fb8d6fd7cb0c5a9f23241d99965bf53b305ed5084d8fcc503bfc24e0"} Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.572863 4677 scope.go:117] "RemoveContainer" containerID="c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.572615 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbc7m" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.607310 4677 scope.go:117] "RemoveContainer" containerID="24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.638613 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nbc7m"] Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.642759 4677 scope.go:117] "RemoveContainer" containerID="83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.649244 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nbc7m"] Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.689472 4677 scope.go:117] "RemoveContainer" containerID="c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8" Dec 03 14:24:08 crc kubenswrapper[4677]: E1203 14:24:08.689780 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8\": container with ID starting with c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8 not found: ID does not exist" containerID="c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.689915 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8"} err="failed to get container status \"c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8\": rpc error: code = NotFound desc = could not find container \"c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8\": container with ID starting with c67f8d1e4d0ba3c3bf57ed2e9dfb0089545133f185489039906ea501563adce8 not found: ID does not exist" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.690007 4677 scope.go:117] "RemoveContainer" containerID="24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8" Dec 03 14:24:08 crc kubenswrapper[4677]: E1203 14:24:08.690308 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8\": container with ID starting with 24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8 not found: ID does not exist" containerID="24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.690377 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8"} err="failed to get container status \"24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8\": rpc error: code = NotFound desc = could not find container \"24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8\": container with ID starting with 24dfb1914192adfef23d8a905d30dcbd3e7ea01246d8e13cbc24b6d19858f9b8 not found: ID does not exist" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.690446 4677 scope.go:117] "RemoveContainer" containerID="83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f" Dec 03 14:24:08 crc kubenswrapper[4677]: E1203 14:24:08.690807 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f\": container with ID starting with 83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f not found: ID does not exist" containerID="83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f" Dec 03 14:24:08 crc kubenswrapper[4677]: I1203 14:24:08.690854 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f"} err="failed to get container status \"83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f\": rpc error: code = NotFound desc = could not find container \"83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f\": container with ID starting with 83b3a6d970d9fa69d760bab5b52f1a9e996d89104ee4f39c15659355edb70c2f not found: ID does not exist" Dec 03 14:24:09 crc kubenswrapper[4677]: I1203 14:24:09.987790 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" path="/var/lib/kubelet/pods/701f2ae9-c397-4e5e-a340-b2abc769990c/volumes" Dec 03 14:24:38 crc kubenswrapper[4677]: I1203 14:24:38.437181 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:24:38 crc kubenswrapper[4677]: I1203 14:24:38.437863 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:24:48 crc kubenswrapper[4677]: I1203 14:24:48.939389 4677 generic.go:334] "Generic (PLEG): container finished" podID="5c6dfb73-f46a-4233-a1bf-a22c6e9e923e" containerID="44e04b3fe97061d4910405e3133dd02b8d1f7f6561b47082475b73c1ba989f8d" exitCode=0 Dec 03 14:24:48 crc kubenswrapper[4677]: I1203 14:24:48.939487 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" event={"ID":"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e","Type":"ContainerDied","Data":"44e04b3fe97061d4910405e3133dd02b8d1f7f6561b47082475b73c1ba989f8d"} Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.372585 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.568620 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-inventory\") pod \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.568925 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-ssh-key\") pod \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.568975 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9xmp\" (UniqueName: \"kubernetes.io/projected/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-kube-api-access-j9xmp\") pod \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\" (UID: \"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e\") " Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.574346 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-kube-api-access-j9xmp" (OuterVolumeSpecName: "kube-api-access-j9xmp") pod "5c6dfb73-f46a-4233-a1bf-a22c6e9e923e" (UID: "5c6dfb73-f46a-4233-a1bf-a22c6e9e923e"). InnerVolumeSpecName "kube-api-access-j9xmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.599450 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-inventory" (OuterVolumeSpecName: "inventory") pod "5c6dfb73-f46a-4233-a1bf-a22c6e9e923e" (UID: "5c6dfb73-f46a-4233-a1bf-a22c6e9e923e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.602860 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5c6dfb73-f46a-4233-a1bf-a22c6e9e923e" (UID: "5c6dfb73-f46a-4233-a1bf-a22c6e9e923e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.672137 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.672175 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.672188 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9xmp\" (UniqueName: \"kubernetes.io/projected/5c6dfb73-f46a-4233-a1bf-a22c6e9e923e-kube-api-access-j9xmp\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.957605 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" event={"ID":"5c6dfb73-f46a-4233-a1bf-a22c6e9e923e","Type":"ContainerDied","Data":"a79bf165a3c661c212bd2c65182096b18af1e5078f304ffe6faefb592f2369ca"} Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.957665 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a79bf165a3c661c212bd2c65182096b18af1e5078f304ffe6faefb592f2369ca" Dec 03 14:24:50 crc kubenswrapper[4677]: I1203 14:24:50.957669 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5q8k" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.068692 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs"] Dec 03 14:24:51 crc kubenswrapper[4677]: E1203 14:24:51.069274 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="extract-content" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.069305 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="extract-content" Dec 03 14:24:51 crc kubenswrapper[4677]: E1203 14:24:51.069332 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6dfb73-f46a-4233-a1bf-a22c6e9e923e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.069344 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6dfb73-f46a-4233-a1bf-a22c6e9e923e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:24:51 crc kubenswrapper[4677]: E1203 14:24:51.069383 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="extract-utilities" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.069392 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="extract-utilities" Dec 03 14:24:51 crc kubenswrapper[4677]: E1203 14:24:51.069412 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="registry-server" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.069420 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="registry-server" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.069716 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c6dfb73-f46a-4233-a1bf-a22c6e9e923e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.069740 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="701f2ae9-c397-4e5e-a340-b2abc769990c" containerName="registry-server" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.070719 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.074331 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.074574 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.074739 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.075908 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.082368 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs"] Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.181644 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.181708 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67x5g\" (UniqueName: \"kubernetes.io/projected/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-kube-api-access-67x5g\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.181762 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.284333 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.284391 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67x5g\" (UniqueName: \"kubernetes.io/projected/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-kube-api-access-67x5g\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.284460 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.290057 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.295125 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.304598 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67x5g\" (UniqueName: \"kubernetes.io/projected/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-kube-api-access-67x5g\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.398914 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.912793 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs"] Dec 03 14:24:51 crc kubenswrapper[4677]: W1203 14:24:51.916193 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3731b0f_16ad_4e49_8a6b_e1d20a118e86.slice/crio-01862f5a27689638aad5abf8324993031eabdb2fe5573db10e34ffcbb519efba WatchSource:0}: Error finding container 01862f5a27689638aad5abf8324993031eabdb2fe5573db10e34ffcbb519efba: Status 404 returned error can't find the container with id 01862f5a27689638aad5abf8324993031eabdb2fe5573db10e34ffcbb519efba Dec 03 14:24:51 crc kubenswrapper[4677]: I1203 14:24:51.966637 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" event={"ID":"b3731b0f-16ad-4e49-8a6b-e1d20a118e86","Type":"ContainerStarted","Data":"01862f5a27689638aad5abf8324993031eabdb2fe5573db10e34ffcbb519efba"} Dec 03 14:24:56 crc kubenswrapper[4677]: I1203 14:24:56.007718 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" event={"ID":"b3731b0f-16ad-4e49-8a6b-e1d20a118e86","Type":"ContainerStarted","Data":"81f3485fc38fd742ac3a45eccb651fd8a3bfa86f209f3790e23a9a37bda706c5"} Dec 03 14:24:56 crc kubenswrapper[4677]: I1203 14:24:56.039846 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" podStartSLOduration=1.7476522719999998 podStartE2EDuration="5.039816649s" podCreationTimestamp="2025-12-03 14:24:51 +0000 UTC" firstStartedPulling="2025-12-03 14:24:51.918985868 +0000 UTC m=+2282.665318323" lastFinishedPulling="2025-12-03 14:24:55.211150245 +0000 UTC m=+2285.957482700" observedRunningTime="2025-12-03 14:24:56.038122682 +0000 UTC m=+2286.784455137" watchObservedRunningTime="2025-12-03 14:24:56.039816649 +0000 UTC m=+2286.786149124" Dec 03 14:25:08 crc kubenswrapper[4677]: I1203 14:25:08.437300 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:25:08 crc kubenswrapper[4677]: I1203 14:25:08.438042 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:25:38 crc kubenswrapper[4677]: I1203 14:25:38.436842 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:25:38 crc kubenswrapper[4677]: I1203 14:25:38.437416 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:25:38 crc kubenswrapper[4677]: I1203 14:25:38.437472 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:25:38 crc kubenswrapper[4677]: I1203 14:25:38.438296 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:25:38 crc kubenswrapper[4677]: I1203 14:25:38.438366 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" gracePeriod=600 Dec 03 14:25:38 crc kubenswrapper[4677]: E1203 14:25:38.565238 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:25:39 crc kubenswrapper[4677]: I1203 14:25:39.431093 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" exitCode=0 Dec 03 14:25:39 crc kubenswrapper[4677]: I1203 14:25:39.431175 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e"} Dec 03 14:25:39 crc kubenswrapper[4677]: I1203 14:25:39.431571 4677 scope.go:117] "RemoveContainer" containerID="dc95799bd3832da5bee136d530caa6131c382dcae6b9ac4b24eb38e3ea0e1e21" Dec 03 14:25:39 crc kubenswrapper[4677]: I1203 14:25:39.433277 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:25:39 crc kubenswrapper[4677]: E1203 14:25:39.433725 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:25:50 crc kubenswrapper[4677]: I1203 14:25:50.538311 4677 generic.go:334] "Generic (PLEG): container finished" podID="b3731b0f-16ad-4e49-8a6b-e1d20a118e86" containerID="81f3485fc38fd742ac3a45eccb651fd8a3bfa86f209f3790e23a9a37bda706c5" exitCode=0 Dec 03 14:25:50 crc kubenswrapper[4677]: I1203 14:25:50.538418 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" event={"ID":"b3731b0f-16ad-4e49-8a6b-e1d20a118e86","Type":"ContainerDied","Data":"81f3485fc38fd742ac3a45eccb651fd8a3bfa86f209f3790e23a9a37bda706c5"} Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.003937 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.110317 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67x5g\" (UniqueName: \"kubernetes.io/projected/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-kube-api-access-67x5g\") pod \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.110397 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-inventory\") pod \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.110479 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-ssh-key\") pod \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\" (UID: \"b3731b0f-16ad-4e49-8a6b-e1d20a118e86\") " Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.117159 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-kube-api-access-67x5g" (OuterVolumeSpecName: "kube-api-access-67x5g") pod "b3731b0f-16ad-4e49-8a6b-e1d20a118e86" (UID: "b3731b0f-16ad-4e49-8a6b-e1d20a118e86"). InnerVolumeSpecName "kube-api-access-67x5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.140332 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-inventory" (OuterVolumeSpecName: "inventory") pod "b3731b0f-16ad-4e49-8a6b-e1d20a118e86" (UID: "b3731b0f-16ad-4e49-8a6b-e1d20a118e86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.146140 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3731b0f-16ad-4e49-8a6b-e1d20a118e86" (UID: "b3731b0f-16ad-4e49-8a6b-e1d20a118e86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.213130 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67x5g\" (UniqueName: \"kubernetes.io/projected/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-kube-api-access-67x5g\") on node \"crc\" DevicePath \"\"" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.213444 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.213455 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3731b0f-16ad-4e49-8a6b-e1d20a118e86-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.563150 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" event={"ID":"b3731b0f-16ad-4e49-8a6b-e1d20a118e86","Type":"ContainerDied","Data":"01862f5a27689638aad5abf8324993031eabdb2fe5573db10e34ffcbb519efba"} Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.563236 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01862f5a27689638aad5abf8324993031eabdb2fe5573db10e34ffcbb519efba" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.563345 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.656884 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pwcrc"] Dec 03 14:25:52 crc kubenswrapper[4677]: E1203 14:25:52.657437 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3731b0f-16ad-4e49-8a6b-e1d20a118e86" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.657458 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3731b0f-16ad-4e49-8a6b-e1d20a118e86" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.657679 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3731b0f-16ad-4e49-8a6b-e1d20a118e86" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.658443 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.660412 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.660694 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.661091 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.661218 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.666503 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pwcrc"] Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.825897 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plnft\" (UniqueName: \"kubernetes.io/projected/da805608-f665-4c50-baf6-b90b94c6fd2c-kube-api-access-plnft\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.826012 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.826085 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.927404 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plnft\" (UniqueName: \"kubernetes.io/projected/da805608-f665-4c50-baf6-b90b94c6fd2c-kube-api-access-plnft\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.927490 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.927570 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.931940 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.932267 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.945875 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plnft\" (UniqueName: \"kubernetes.io/projected/da805608-f665-4c50-baf6-b90b94c6fd2c-kube-api-access-plnft\") pod \"ssh-known-hosts-edpm-deployment-pwcrc\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:52 crc kubenswrapper[4677]: I1203 14:25:52.980454 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:25:53 crc kubenswrapper[4677]: I1203 14:25:53.488620 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pwcrc"] Dec 03 14:25:53 crc kubenswrapper[4677]: I1203 14:25:53.572533 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" event={"ID":"da805608-f665-4c50-baf6-b90b94c6fd2c","Type":"ContainerStarted","Data":"dbfe37fb78b4edc803e99931dd9379552b5080b5c87ff5b20f75eeee6cc3ffe7"} Dec 03 14:25:53 crc kubenswrapper[4677]: I1203 14:25:53.977843 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:25:53 crc kubenswrapper[4677]: E1203 14:25:53.978216 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:25:55 crc kubenswrapper[4677]: I1203 14:25:55.601084 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" event={"ID":"da805608-f665-4c50-baf6-b90b94c6fd2c","Type":"ContainerStarted","Data":"01a91b70aa5e76ac75fe880d4c53f30fa5e1926ce97c03d1021d68791ef4ed34"} Dec 03 14:25:55 crc kubenswrapper[4677]: I1203 14:25:55.627544 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" podStartSLOduration=2.4588672750000002 podStartE2EDuration="3.627528561s" podCreationTimestamp="2025-12-03 14:25:52 +0000 UTC" firstStartedPulling="2025-12-03 14:25:53.500516357 +0000 UTC m=+2344.246848812" lastFinishedPulling="2025-12-03 14:25:54.669177643 +0000 UTC m=+2345.415510098" observedRunningTime="2025-12-03 14:25:55.624167967 +0000 UTC m=+2346.370500442" watchObservedRunningTime="2025-12-03 14:25:55.627528561 +0000 UTC m=+2346.373861016" Dec 03 14:26:02 crc kubenswrapper[4677]: I1203 14:26:02.661894 4677 generic.go:334] "Generic (PLEG): container finished" podID="da805608-f665-4c50-baf6-b90b94c6fd2c" containerID="01a91b70aa5e76ac75fe880d4c53f30fa5e1926ce97c03d1021d68791ef4ed34" exitCode=0 Dec 03 14:26:02 crc kubenswrapper[4677]: I1203 14:26:02.662154 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" event={"ID":"da805608-f665-4c50-baf6-b90b94c6fd2c","Type":"ContainerDied","Data":"01a91b70aa5e76ac75fe880d4c53f30fa5e1926ce97c03d1021d68791ef4ed34"} Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.082901 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.244874 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-inventory-0\") pod \"da805608-f665-4c50-baf6-b90b94c6fd2c\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.245130 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-ssh-key-openstack-edpm-ipam\") pod \"da805608-f665-4c50-baf6-b90b94c6fd2c\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.245207 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plnft\" (UniqueName: \"kubernetes.io/projected/da805608-f665-4c50-baf6-b90b94c6fd2c-kube-api-access-plnft\") pod \"da805608-f665-4c50-baf6-b90b94c6fd2c\" (UID: \"da805608-f665-4c50-baf6-b90b94c6fd2c\") " Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.251074 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da805608-f665-4c50-baf6-b90b94c6fd2c-kube-api-access-plnft" (OuterVolumeSpecName: "kube-api-access-plnft") pod "da805608-f665-4c50-baf6-b90b94c6fd2c" (UID: "da805608-f665-4c50-baf6-b90b94c6fd2c"). InnerVolumeSpecName "kube-api-access-plnft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.276706 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "da805608-f665-4c50-baf6-b90b94c6fd2c" (UID: "da805608-f665-4c50-baf6-b90b94c6fd2c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.277141 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "da805608-f665-4c50-baf6-b90b94c6fd2c" (UID: "da805608-f665-4c50-baf6-b90b94c6fd2c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.347665 4677 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.348064 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da805608-f665-4c50-baf6-b90b94c6fd2c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.348215 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plnft\" (UniqueName: \"kubernetes.io/projected/da805608-f665-4c50-baf6-b90b94c6fd2c-kube-api-access-plnft\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.681791 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.681842 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pwcrc" event={"ID":"da805608-f665-4c50-baf6-b90b94c6fd2c","Type":"ContainerDied","Data":"dbfe37fb78b4edc803e99931dd9379552b5080b5c87ff5b20f75eeee6cc3ffe7"} Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.681880 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbfe37fb78b4edc803e99931dd9379552b5080b5c87ff5b20f75eeee6cc3ffe7" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.769827 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh"] Dec 03 14:26:04 crc kubenswrapper[4677]: E1203 14:26:04.770637 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da805608-f665-4c50-baf6-b90b94c6fd2c" containerName="ssh-known-hosts-edpm-deployment" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.770730 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="da805608-f665-4c50-baf6-b90b94c6fd2c" containerName="ssh-known-hosts-edpm-deployment" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.771087 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="da805608-f665-4c50-baf6-b90b94c6fd2c" containerName="ssh-known-hosts-edpm-deployment" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.771964 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.773787 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.774627 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.776149 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.776329 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.782618 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh"] Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.959882 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.960054 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:04 crc kubenswrapper[4677]: I1203 14:26:04.960082 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn4wc\" (UniqueName: \"kubernetes.io/projected/5f602efb-8c0c-4ba7-a297-d24706833f00-kube-api-access-kn4wc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.062190 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.062351 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.062390 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn4wc\" (UniqueName: \"kubernetes.io/projected/5f602efb-8c0c-4ba7-a297-d24706833f00-kube-api-access-kn4wc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.066143 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.068649 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.083866 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn4wc\" (UniqueName: \"kubernetes.io/projected/5f602efb-8c0c-4ba7-a297-d24706833f00-kube-api-access-kn4wc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kx6wh\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.089017 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.623223 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh"] Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.631815 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:26:05 crc kubenswrapper[4677]: I1203 14:26:05.691328 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" event={"ID":"5f602efb-8c0c-4ba7-a297-d24706833f00","Type":"ContainerStarted","Data":"b6907a36ee8511945f683ecce731c6feb92f1d31583a187caea8c70bb6e83718"} Dec 03 14:26:06 crc kubenswrapper[4677]: I1203 14:26:06.701024 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" event={"ID":"5f602efb-8c0c-4ba7-a297-d24706833f00","Type":"ContainerStarted","Data":"7b0a609a7c8d6698b564560441e76d1b9b46da67c16699e09fba91ef506a0729"} Dec 03 14:26:06 crc kubenswrapper[4677]: I1203 14:26:06.723536 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" podStartSLOduration=1.921964714 podStartE2EDuration="2.723518858s" podCreationTimestamp="2025-12-03 14:26:04 +0000 UTC" firstStartedPulling="2025-12-03 14:26:05.631238203 +0000 UTC m=+2356.377570668" lastFinishedPulling="2025-12-03 14:26:06.432792357 +0000 UTC m=+2357.179124812" observedRunningTime="2025-12-03 14:26:06.71668107 +0000 UTC m=+2357.463013535" watchObservedRunningTime="2025-12-03 14:26:06.723518858 +0000 UTC m=+2357.469851313" Dec 03 14:26:06 crc kubenswrapper[4677]: I1203 14:26:06.976777 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:26:06 crc kubenswrapper[4677]: E1203 14:26:06.977682 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:26:15 crc kubenswrapper[4677]: I1203 14:26:15.784287 4677 generic.go:334] "Generic (PLEG): container finished" podID="5f602efb-8c0c-4ba7-a297-d24706833f00" containerID="7b0a609a7c8d6698b564560441e76d1b9b46da67c16699e09fba91ef506a0729" exitCode=0 Dec 03 14:26:15 crc kubenswrapper[4677]: I1203 14:26:15.784327 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" event={"ID":"5f602efb-8c0c-4ba7-a297-d24706833f00","Type":"ContainerDied","Data":"7b0a609a7c8d6698b564560441e76d1b9b46da67c16699e09fba91ef506a0729"} Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.218463 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.302452 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kn4wc\" (UniqueName: \"kubernetes.io/projected/5f602efb-8c0c-4ba7-a297-d24706833f00-kube-api-access-kn4wc\") pod \"5f602efb-8c0c-4ba7-a297-d24706833f00\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.303145 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-inventory\") pod \"5f602efb-8c0c-4ba7-a297-d24706833f00\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.303213 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-ssh-key\") pod \"5f602efb-8c0c-4ba7-a297-d24706833f00\" (UID: \"5f602efb-8c0c-4ba7-a297-d24706833f00\") " Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.312146 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f602efb-8c0c-4ba7-a297-d24706833f00-kube-api-access-kn4wc" (OuterVolumeSpecName: "kube-api-access-kn4wc") pod "5f602efb-8c0c-4ba7-a297-d24706833f00" (UID: "5f602efb-8c0c-4ba7-a297-d24706833f00"). InnerVolumeSpecName "kube-api-access-kn4wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.334656 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f602efb-8c0c-4ba7-a297-d24706833f00" (UID: "5f602efb-8c0c-4ba7-a297-d24706833f00"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.340773 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-inventory" (OuterVolumeSpecName: "inventory") pod "5f602efb-8c0c-4ba7-a297-d24706833f00" (UID: "5f602efb-8c0c-4ba7-a297-d24706833f00"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.405925 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kn4wc\" (UniqueName: \"kubernetes.io/projected/5f602efb-8c0c-4ba7-a297-d24706833f00-kube-api-access-kn4wc\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.405986 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.405999 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f602efb-8c0c-4ba7-a297-d24706833f00-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.803998 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" event={"ID":"5f602efb-8c0c-4ba7-a297-d24706833f00","Type":"ContainerDied","Data":"b6907a36ee8511945f683ecce731c6feb92f1d31583a187caea8c70bb6e83718"} Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.804050 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6907a36ee8511945f683ecce731c6feb92f1d31583a187caea8c70bb6e83718" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.804052 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kx6wh" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.936886 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg"] Dec 03 14:26:17 crc kubenswrapper[4677]: E1203 14:26:17.937389 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f602efb-8c0c-4ba7-a297-d24706833f00" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.937414 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f602efb-8c0c-4ba7-a297-d24706833f00" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.937655 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f602efb-8c0c-4ba7-a297-d24706833f00" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.938519 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.940985 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.941303 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.941628 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.945576 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:26:17 crc kubenswrapper[4677]: I1203 14:26:17.950221 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg"] Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.016880 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.017097 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.017152 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s5qf\" (UniqueName: \"kubernetes.io/projected/054f88ba-5afa-48e7-89c5-0573ae4dc14b-kube-api-access-2s5qf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.119827 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.119982 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.120024 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s5qf\" (UniqueName: \"kubernetes.io/projected/054f88ba-5afa-48e7-89c5-0573ae4dc14b-kube-api-access-2s5qf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.123780 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.124590 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.138012 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s5qf\" (UniqueName: \"kubernetes.io/projected/054f88ba-5afa-48e7-89c5-0573ae4dc14b-kube-api-access-2s5qf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.261594 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.779631 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg"] Dec 03 14:26:18 crc kubenswrapper[4677]: I1203 14:26:18.846031 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" event={"ID":"054f88ba-5afa-48e7-89c5-0573ae4dc14b","Type":"ContainerStarted","Data":"81ad7d064c3f8a6e3d123a26136e9fd70307e4ac6883ef0163845cbc35418654"} Dec 03 14:26:20 crc kubenswrapper[4677]: I1203 14:26:20.870136 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" event={"ID":"054f88ba-5afa-48e7-89c5-0573ae4dc14b","Type":"ContainerStarted","Data":"2497443500df0982afb1d8527ea23a80e1f9fab0ea9fd3426745eac4ea629ba3"} Dec 03 14:26:20 crc kubenswrapper[4677]: I1203 14:26:20.888882 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" podStartSLOduration=3.4332033539999998 podStartE2EDuration="3.888863533s" podCreationTimestamp="2025-12-03 14:26:17 +0000 UTC" firstStartedPulling="2025-12-03 14:26:18.796736103 +0000 UTC m=+2369.543068558" lastFinishedPulling="2025-12-03 14:26:19.252396272 +0000 UTC m=+2369.998728737" observedRunningTime="2025-12-03 14:26:20.885197812 +0000 UTC m=+2371.631530287" watchObservedRunningTime="2025-12-03 14:26:20.888863533 +0000 UTC m=+2371.635195988" Dec 03 14:26:21 crc kubenswrapper[4677]: I1203 14:26:21.976832 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:26:21 crc kubenswrapper[4677]: E1203 14:26:21.977435 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:26:30 crc kubenswrapper[4677]: I1203 14:26:30.955464 4677 generic.go:334] "Generic (PLEG): container finished" podID="054f88ba-5afa-48e7-89c5-0573ae4dc14b" containerID="2497443500df0982afb1d8527ea23a80e1f9fab0ea9fd3426745eac4ea629ba3" exitCode=0 Dec 03 14:26:30 crc kubenswrapper[4677]: I1203 14:26:30.955572 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" event={"ID":"054f88ba-5afa-48e7-89c5-0573ae4dc14b","Type":"ContainerDied","Data":"2497443500df0982afb1d8527ea23a80e1f9fab0ea9fd3426745eac4ea629ba3"} Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.377145 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.506293 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s5qf\" (UniqueName: \"kubernetes.io/projected/054f88ba-5afa-48e7-89c5-0573ae4dc14b-kube-api-access-2s5qf\") pod \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.506481 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-ssh-key\") pod \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.506646 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-inventory\") pod \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\" (UID: \"054f88ba-5afa-48e7-89c5-0573ae4dc14b\") " Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.512870 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/054f88ba-5afa-48e7-89c5-0573ae4dc14b-kube-api-access-2s5qf" (OuterVolumeSpecName: "kube-api-access-2s5qf") pod "054f88ba-5afa-48e7-89c5-0573ae4dc14b" (UID: "054f88ba-5afa-48e7-89c5-0573ae4dc14b"). InnerVolumeSpecName "kube-api-access-2s5qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.538505 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "054f88ba-5afa-48e7-89c5-0573ae4dc14b" (UID: "054f88ba-5afa-48e7-89c5-0573ae4dc14b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.539923 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-inventory" (OuterVolumeSpecName: "inventory") pod "054f88ba-5afa-48e7-89c5-0573ae4dc14b" (UID: "054f88ba-5afa-48e7-89c5-0573ae4dc14b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.608549 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s5qf\" (UniqueName: \"kubernetes.io/projected/054f88ba-5afa-48e7-89c5-0573ae4dc14b-kube-api-access-2s5qf\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.608898 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.608909 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054f88ba-5afa-48e7-89c5-0573ae4dc14b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.975814 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" event={"ID":"054f88ba-5afa-48e7-89c5-0573ae4dc14b","Type":"ContainerDied","Data":"81ad7d064c3f8a6e3d123a26136e9fd70307e4ac6883ef0163845cbc35418654"} Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.975875 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81ad7d064c3f8a6e3d123a26136e9fd70307e4ac6883ef0163845cbc35418654" Dec 03 14:26:32 crc kubenswrapper[4677]: I1203 14:26:32.976607 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg" Dec 03 14:26:33 crc kubenswrapper[4677]: E1203 14:26:33.127585 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod054f88ba_5afa_48e7_89c5_0573ae4dc14b.slice/crio-81ad7d064c3f8a6e3d123a26136e9fd70307e4ac6883ef0163845cbc35418654\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod054f88ba_5afa_48e7_89c5_0573ae4dc14b.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.139086 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d"] Dec 03 14:26:33 crc kubenswrapper[4677]: E1203 14:26:33.139690 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054f88ba-5afa-48e7-89c5-0573ae4dc14b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.139733 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="054f88ba-5afa-48e7-89c5-0573ae4dc14b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.140046 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="054f88ba-5afa-48e7-89c5-0573ae4dc14b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.140906 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.145168 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.145243 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.145518 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.145172 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.145773 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.145879 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.150386 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.150691 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.168276 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d"] Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.338081 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.338807 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.338847 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339055 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339114 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339213 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339266 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339302 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339331 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339463 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339540 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22knq\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-kube-api-access-22knq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339593 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339736 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.339840 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.441751 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442657 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442731 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442777 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442803 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442837 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442863 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442910 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442943 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.442984 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.443012 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.443039 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.443063 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22knq\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-kube-api-access-22knq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.443084 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.447215 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.449242 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.449685 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.450679 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.451380 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.451439 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.451681 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.451906 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.452878 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.453100 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.453241 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.453378 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.453454 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.463685 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22knq\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-kube-api-access-22knq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:33 crc kubenswrapper[4677]: I1203 14:26:33.522182 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:26:34 crc kubenswrapper[4677]: I1203 14:26:34.068511 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d"] Dec 03 14:26:34 crc kubenswrapper[4677]: I1203 14:26:34.975741 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:26:34 crc kubenswrapper[4677]: E1203 14:26:34.976443 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:26:35 crc kubenswrapper[4677]: I1203 14:26:35.002373 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" event={"ID":"b1374bac-1ae5-47db-8fc5-736d874a8a55","Type":"ContainerStarted","Data":"2a3297853e4f6f3cd13808497a5f9a9bea2927d092e2e88a03e2df4277148625"} Dec 03 14:26:35 crc kubenswrapper[4677]: I1203 14:26:35.002426 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" event={"ID":"b1374bac-1ae5-47db-8fc5-736d874a8a55","Type":"ContainerStarted","Data":"e31a151fcd0527da57c9eb27486f298769fa3361366cf47e82d36f8535db0c7b"} Dec 03 14:26:35 crc kubenswrapper[4677]: I1203 14:26:35.029545 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" podStartSLOduration=1.515554737 podStartE2EDuration="2.029522766s" podCreationTimestamp="2025-12-03 14:26:33 +0000 UTC" firstStartedPulling="2025-12-03 14:26:34.071475078 +0000 UTC m=+2384.817807533" lastFinishedPulling="2025-12-03 14:26:34.585443107 +0000 UTC m=+2385.331775562" observedRunningTime="2025-12-03 14:26:35.025310549 +0000 UTC m=+2385.771643014" watchObservedRunningTime="2025-12-03 14:26:35.029522766 +0000 UTC m=+2385.775855221" Dec 03 14:26:45 crc kubenswrapper[4677]: I1203 14:26:45.976470 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:26:45 crc kubenswrapper[4677]: E1203 14:26:45.977217 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:26:56 crc kubenswrapper[4677]: I1203 14:26:56.977222 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:26:56 crc kubenswrapper[4677]: E1203 14:26:56.978231 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:27:10 crc kubenswrapper[4677]: I1203 14:27:10.976853 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:27:10 crc kubenswrapper[4677]: E1203 14:27:10.977665 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:27:14 crc kubenswrapper[4677]: I1203 14:27:14.376390 4677 generic.go:334] "Generic (PLEG): container finished" podID="b1374bac-1ae5-47db-8fc5-736d874a8a55" containerID="2a3297853e4f6f3cd13808497a5f9a9bea2927d092e2e88a03e2df4277148625" exitCode=0 Dec 03 14:27:14 crc kubenswrapper[4677]: I1203 14:27:14.376493 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" event={"ID":"b1374bac-1ae5-47db-8fc5-736d874a8a55","Type":"ContainerDied","Data":"2a3297853e4f6f3cd13808497a5f9a9bea2927d092e2e88a03e2df4277148625"} Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.779452 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.895299 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-repo-setup-combined-ca-bundle\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.895400 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-libvirt-combined-ca-bundle\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.895492 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ovn-combined-ca-bundle\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.895541 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.895639 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-neutron-metadata-combined-ca-bundle\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.895723 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.896434 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ssh-key\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.896563 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.896640 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.896747 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-bootstrap-combined-ca-bundle\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.896795 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22knq\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-kube-api-access-22knq\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.896856 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-telemetry-combined-ca-bundle\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.896908 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-nova-combined-ca-bundle\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.897045 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-inventory\") pod \"b1374bac-1ae5-47db-8fc5-736d874a8a55\" (UID: \"b1374bac-1ae5-47db-8fc5-736d874a8a55\") " Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.902910 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.902910 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.903359 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.904344 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.904392 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.904363 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.905194 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.905698 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-kube-api-access-22knq" (OuterVolumeSpecName: "kube-api-access-22knq") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "kube-api-access-22knq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.906327 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.907295 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.907363 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.907713 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.931050 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:15 crc kubenswrapper[4677]: I1203 14:27:15.942857 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-inventory" (OuterVolumeSpecName: "inventory") pod "b1374bac-1ae5-47db-8fc5-736d874a8a55" (UID: "b1374bac-1ae5-47db-8fc5-736d874a8a55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000078 4677 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000113 4677 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000126 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000138 4677 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000148 4677 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000158 4677 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000180 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22knq\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-kube-api-access-22knq\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000189 4677 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000197 4677 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000206 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000214 4677 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000224 4677 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000234 4677 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1374bac-1ae5-47db-8fc5-736d874a8a55-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.000242 4677 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b1374bac-1ae5-47db-8fc5-736d874a8a55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.394715 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" event={"ID":"b1374bac-1ae5-47db-8fc5-736d874a8a55","Type":"ContainerDied","Data":"e31a151fcd0527da57c9eb27486f298769fa3361366cf47e82d36f8535db0c7b"} Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.394990 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e31a151fcd0527da57c9eb27486f298769fa3361366cf47e82d36f8535db0c7b" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.395017 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.520919 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf"] Dec 03 14:27:16 crc kubenswrapper[4677]: E1203 14:27:16.521761 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1374bac-1ae5-47db-8fc5-736d874a8a55" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.521788 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1374bac-1ae5-47db-8fc5-736d874a8a55" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.522075 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1374bac-1ae5-47db-8fc5-736d874a8a55" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.522887 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.525938 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.526004 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.528630 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.528857 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.529822 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.530821 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf"] Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.697265 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.698356 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n99h9\" (UniqueName: \"kubernetes.io/projected/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-kube-api-access-n99h9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.698601 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.698848 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.699323 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.800763 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.800829 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.800916 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.801005 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.801092 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n99h9\" (UniqueName: \"kubernetes.io/projected/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-kube-api-access-n99h9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.801929 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.804805 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.805821 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.806469 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.818853 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n99h9\" (UniqueName: \"kubernetes.io/projected/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-kube-api-access-n99h9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nr8kf\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:16 crc kubenswrapper[4677]: I1203 14:27:16.846454 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:27:17 crc kubenswrapper[4677]: I1203 14:27:17.333124 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf"] Dec 03 14:27:17 crc kubenswrapper[4677]: I1203 14:27:17.404117 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" event={"ID":"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6","Type":"ContainerStarted","Data":"439b7fa0a4699a1b06836f46d6474594efe0248f83ec5b7b6e699b931485bb20"} Dec 03 14:27:18 crc kubenswrapper[4677]: I1203 14:27:18.417123 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" event={"ID":"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6","Type":"ContainerStarted","Data":"94031aab705f4d2708a4ddbaa0cadcf47246939369023d63ba180e87dc306cdc"} Dec 03 14:27:18 crc kubenswrapper[4677]: I1203 14:27:18.445884 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" podStartSLOduration=2.01342691 podStartE2EDuration="2.445858107s" podCreationTimestamp="2025-12-03 14:27:16 +0000 UTC" firstStartedPulling="2025-12-03 14:27:17.337466485 +0000 UTC m=+2428.083798940" lastFinishedPulling="2025-12-03 14:27:17.769897682 +0000 UTC m=+2428.516230137" observedRunningTime="2025-12-03 14:27:18.440108648 +0000 UTC m=+2429.186441103" watchObservedRunningTime="2025-12-03 14:27:18.445858107 +0000 UTC m=+2429.192190562" Dec 03 14:27:23 crc kubenswrapper[4677]: I1203 14:27:23.976693 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:27:23 crc kubenswrapper[4677]: E1203 14:27:23.977674 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:27:35 crc kubenswrapper[4677]: I1203 14:27:35.977354 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:27:35 crc kubenswrapper[4677]: E1203 14:27:35.978418 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:27:46 crc kubenswrapper[4677]: I1203 14:27:46.976935 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:27:46 crc kubenswrapper[4677]: E1203 14:27:46.977696 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:28:00 crc kubenswrapper[4677]: I1203 14:28:00.976314 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:28:00 crc kubenswrapper[4677]: E1203 14:28:00.977082 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:28:11 crc kubenswrapper[4677]: I1203 14:28:11.976287 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:28:11 crc kubenswrapper[4677]: E1203 14:28:11.977306 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:28:21 crc kubenswrapper[4677]: I1203 14:28:21.975455 4677 generic.go:334] "Generic (PLEG): container finished" podID="aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" containerID="94031aab705f4d2708a4ddbaa0cadcf47246939369023d63ba180e87dc306cdc" exitCode=0 Dec 03 14:28:21 crc kubenswrapper[4677]: I1203 14:28:21.986072 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" event={"ID":"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6","Type":"ContainerDied","Data":"94031aab705f4d2708a4ddbaa0cadcf47246939369023d63ba180e87dc306cdc"} Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.386480 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.475431 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovn-combined-ca-bundle\") pod \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.475470 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovncontroller-config-0\") pod \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.475567 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ssh-key\") pod \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.475630 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-inventory\") pod \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.475727 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n99h9\" (UniqueName: \"kubernetes.io/projected/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-kube-api-access-n99h9\") pod \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\" (UID: \"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6\") " Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.480977 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" (UID: "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.482131 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-kube-api-access-n99h9" (OuterVolumeSpecName: "kube-api-access-n99h9") pod "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" (UID: "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6"). InnerVolumeSpecName "kube-api-access-n99h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.507045 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" (UID: "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.507469 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" (UID: "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.508813 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-inventory" (OuterVolumeSpecName: "inventory") pod "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" (UID: "aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.578093 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n99h9\" (UniqueName: \"kubernetes.io/projected/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-kube-api-access-n99h9\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.578123 4677 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.578133 4677 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.578141 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.578150 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.995001 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" event={"ID":"aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6","Type":"ContainerDied","Data":"439b7fa0a4699a1b06836f46d6474594efe0248f83ec5b7b6e699b931485bb20"} Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.995368 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="439b7fa0a4699a1b06836f46d6474594efe0248f83ec5b7b6e699b931485bb20" Dec 03 14:28:23 crc kubenswrapper[4677]: I1203 14:28:23.995080 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nr8kf" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.164192 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px"] Dec 03 14:28:24 crc kubenswrapper[4677]: E1203 14:28:24.164730 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.164757 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.165017 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.165882 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.168098 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.168261 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.172351 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.172620 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.173002 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.173827 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.176123 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px"] Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.291607 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.291675 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.291733 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw4k6\" (UniqueName: \"kubernetes.io/projected/164edf8a-8633-4592-b307-ea42bb77cc8d-kube-api-access-mw4k6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.291777 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.291864 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.291990 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.393521 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.393674 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.393722 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.393780 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw4k6\" (UniqueName: \"kubernetes.io/projected/164edf8a-8633-4592-b307-ea42bb77cc8d-kube-api-access-mw4k6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.393827 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.393902 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.398147 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.398424 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.398830 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.399382 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.401735 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.412841 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw4k6\" (UniqueName: \"kubernetes.io/projected/164edf8a-8633-4592-b307-ea42bb77cc8d-kube-api-access-mw4k6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:24 crc kubenswrapper[4677]: I1203 14:28:24.485520 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:28:25 crc kubenswrapper[4677]: I1203 14:28:25.060376 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px"] Dec 03 14:28:25 crc kubenswrapper[4677]: I1203 14:28:25.978436 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:28:25 crc kubenswrapper[4677]: E1203 14:28:25.979082 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:28:26 crc kubenswrapper[4677]: I1203 14:28:26.017753 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" event={"ID":"164edf8a-8633-4592-b307-ea42bb77cc8d","Type":"ContainerStarted","Data":"47b68ec47e584839c0feee075d67031d676a533d3f50851b9cfb99d7ab1f1664"} Dec 03 14:28:26 crc kubenswrapper[4677]: I1203 14:28:26.017810 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" event={"ID":"164edf8a-8633-4592-b307-ea42bb77cc8d","Type":"ContainerStarted","Data":"04ac72d3a63b218033403cbb703b88a99b7f154a692e3a59f022fecd012a5306"} Dec 03 14:28:26 crc kubenswrapper[4677]: I1203 14:28:26.038969 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" podStartSLOduration=1.572732528 podStartE2EDuration="2.038936256s" podCreationTimestamp="2025-12-03 14:28:24 +0000 UTC" firstStartedPulling="2025-12-03 14:28:25.071346755 +0000 UTC m=+2495.817679210" lastFinishedPulling="2025-12-03 14:28:25.537550473 +0000 UTC m=+2496.283882938" observedRunningTime="2025-12-03 14:28:26.036332504 +0000 UTC m=+2496.782664979" watchObservedRunningTime="2025-12-03 14:28:26.038936256 +0000 UTC m=+2496.785268711" Dec 03 14:28:38 crc kubenswrapper[4677]: I1203 14:28:38.976389 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:28:38 crc kubenswrapper[4677]: E1203 14:28:38.978190 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:28:53 crc kubenswrapper[4677]: I1203 14:28:53.976765 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:28:53 crc kubenswrapper[4677]: E1203 14:28:53.977723 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:29:05 crc kubenswrapper[4677]: I1203 14:29:05.976929 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:29:05 crc kubenswrapper[4677]: E1203 14:29:05.978266 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:29:16 crc kubenswrapper[4677]: I1203 14:29:16.502084 4677 generic.go:334] "Generic (PLEG): container finished" podID="164edf8a-8633-4592-b307-ea42bb77cc8d" containerID="47b68ec47e584839c0feee075d67031d676a533d3f50851b9cfb99d7ab1f1664" exitCode=0 Dec 03 14:29:16 crc kubenswrapper[4677]: I1203 14:29:16.502173 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" event={"ID":"164edf8a-8633-4592-b307-ea42bb77cc8d","Type":"ContainerDied","Data":"47b68ec47e584839c0feee075d67031d676a533d3f50851b9cfb99d7ab1f1664"} Dec 03 14:29:17 crc kubenswrapper[4677]: I1203 14:29:17.926988 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.099926 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw4k6\" (UniqueName: \"kubernetes.io/projected/164edf8a-8633-4592-b307-ea42bb77cc8d-kube-api-access-mw4k6\") pod \"164edf8a-8633-4592-b307-ea42bb77cc8d\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.100308 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-inventory\") pod \"164edf8a-8633-4592-b307-ea42bb77cc8d\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.100390 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-metadata-combined-ca-bundle\") pod \"164edf8a-8633-4592-b307-ea42bb77cc8d\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.100426 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-ssh-key\") pod \"164edf8a-8633-4592-b307-ea42bb77cc8d\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.100553 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-nova-metadata-neutron-config-0\") pod \"164edf8a-8633-4592-b307-ea42bb77cc8d\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.100672 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"164edf8a-8633-4592-b307-ea42bb77cc8d\" (UID: \"164edf8a-8633-4592-b307-ea42bb77cc8d\") " Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.106359 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/164edf8a-8633-4592-b307-ea42bb77cc8d-kube-api-access-mw4k6" (OuterVolumeSpecName: "kube-api-access-mw4k6") pod "164edf8a-8633-4592-b307-ea42bb77cc8d" (UID: "164edf8a-8633-4592-b307-ea42bb77cc8d"). InnerVolumeSpecName "kube-api-access-mw4k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.112173 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "164edf8a-8633-4592-b307-ea42bb77cc8d" (UID: "164edf8a-8633-4592-b307-ea42bb77cc8d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.133825 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "164edf8a-8633-4592-b307-ea42bb77cc8d" (UID: "164edf8a-8633-4592-b307-ea42bb77cc8d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.134537 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "164edf8a-8633-4592-b307-ea42bb77cc8d" (UID: "164edf8a-8633-4592-b307-ea42bb77cc8d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.137429 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "164edf8a-8633-4592-b307-ea42bb77cc8d" (UID: "164edf8a-8633-4592-b307-ea42bb77cc8d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.137847 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-inventory" (OuterVolumeSpecName: "inventory") pod "164edf8a-8633-4592-b307-ea42bb77cc8d" (UID: "164edf8a-8633-4592-b307-ea42bb77cc8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.203112 4677 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.203169 4677 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.203186 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw4k6\" (UniqueName: \"kubernetes.io/projected/164edf8a-8633-4592-b307-ea42bb77cc8d-kube-api-access-mw4k6\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.203198 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.203209 4677 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.203222 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/164edf8a-8633-4592-b307-ea42bb77cc8d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.521740 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" event={"ID":"164edf8a-8633-4592-b307-ea42bb77cc8d","Type":"ContainerDied","Data":"04ac72d3a63b218033403cbb703b88a99b7f154a692e3a59f022fecd012a5306"} Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.521810 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.521817 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04ac72d3a63b218033403cbb703b88a99b7f154a692e3a59f022fecd012a5306" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.611791 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6"] Dec 03 14:29:18 crc kubenswrapper[4677]: E1203 14:29:18.614504 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="164edf8a-8633-4592-b307-ea42bb77cc8d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.614533 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="164edf8a-8633-4592-b307-ea42bb77cc8d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.614838 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="164edf8a-8633-4592-b307-ea42bb77cc8d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.615501 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.619437 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.619729 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.619850 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.620018 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.620151 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.630180 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6"] Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.712904 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.712996 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.713203 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.713265 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.713476 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2259\" (UniqueName: \"kubernetes.io/projected/891909bd-1955-436b-8664-79a41b0676e5-kube-api-access-z2259\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.815116 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2259\" (UniqueName: \"kubernetes.io/projected/891909bd-1955-436b-8664-79a41b0676e5-kube-api-access-z2259\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.815174 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.815194 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.815320 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.815359 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.819767 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.820319 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.820916 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.830444 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.831085 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2259\" (UniqueName: \"kubernetes.io/projected/891909bd-1955-436b-8664-79a41b0676e5-kube-api-access-z2259\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-64ds6\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:18 crc kubenswrapper[4677]: I1203 14:29:18.939968 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:29:19 crc kubenswrapper[4677]: I1203 14:29:19.471831 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6"] Dec 03 14:29:19 crc kubenswrapper[4677]: I1203 14:29:19.532710 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" event={"ID":"891909bd-1955-436b-8664-79a41b0676e5","Type":"ContainerStarted","Data":"ed362823d683e1b09b976c0272a8662b9c3e9b1238383114799333cf2167f0d7"} Dec 03 14:29:19 crc kubenswrapper[4677]: I1203 14:29:19.976136 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:29:19 crc kubenswrapper[4677]: E1203 14:29:19.976507 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:29:20 crc kubenswrapper[4677]: I1203 14:29:20.544973 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" event={"ID":"891909bd-1955-436b-8664-79a41b0676e5","Type":"ContainerStarted","Data":"d0941009e801c12ff036b3576e2260a81f782c1e944a4fb3db327474159378af"} Dec 03 14:29:20 crc kubenswrapper[4677]: I1203 14:29:20.570625 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" podStartSLOduration=2.138074687 podStartE2EDuration="2.570604494s" podCreationTimestamp="2025-12-03 14:29:18 +0000 UTC" firstStartedPulling="2025-12-03 14:29:19.492771823 +0000 UTC m=+2550.239104278" lastFinishedPulling="2025-12-03 14:29:19.92530163 +0000 UTC m=+2550.671634085" observedRunningTime="2025-12-03 14:29:20.567475458 +0000 UTC m=+2551.313807943" watchObservedRunningTime="2025-12-03 14:29:20.570604494 +0000 UTC m=+2551.316936949" Dec 03 14:29:33 crc kubenswrapper[4677]: I1203 14:29:33.976076 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:29:33 crc kubenswrapper[4677]: E1203 14:29:33.976864 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:29:46 crc kubenswrapper[4677]: I1203 14:29:46.976067 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:29:46 crc kubenswrapper[4677]: E1203 14:29:46.976896 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:29:57 crc kubenswrapper[4677]: I1203 14:29:57.976143 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:29:57 crc kubenswrapper[4677]: E1203 14:29:57.977046 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.163857 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms"] Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.167094 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.169080 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.169255 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.178178 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms"] Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.261966 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef734bef-9a2b-4622-b8e8-d476cebd82ee-secret-volume\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.262088 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef734bef-9a2b-4622-b8e8-d476cebd82ee-config-volume\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.262173 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xggmb\" (UniqueName: \"kubernetes.io/projected/ef734bef-9a2b-4622-b8e8-d476cebd82ee-kube-api-access-xggmb\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.363928 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xggmb\" (UniqueName: \"kubernetes.io/projected/ef734bef-9a2b-4622-b8e8-d476cebd82ee-kube-api-access-xggmb\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.364742 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef734bef-9a2b-4622-b8e8-d476cebd82ee-secret-volume\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.364915 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef734bef-9a2b-4622-b8e8-d476cebd82ee-config-volume\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.365930 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef734bef-9a2b-4622-b8e8-d476cebd82ee-config-volume\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.371543 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef734bef-9a2b-4622-b8e8-d476cebd82ee-secret-volume\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.385716 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xggmb\" (UniqueName: \"kubernetes.io/projected/ef734bef-9a2b-4622-b8e8-d476cebd82ee-kube-api-access-xggmb\") pod \"collect-profiles-29412870-2tbms\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.502139 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:00 crc kubenswrapper[4677]: I1203 14:30:00.969814 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms"] Dec 03 14:30:01 crc kubenswrapper[4677]: I1203 14:30:01.939200 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef734bef-9a2b-4622-b8e8-d476cebd82ee" containerID="325ba1ba74b7328ef77ee4badd32e67dbe7042d8c02ab66efa27598016342674" exitCode=0 Dec 03 14:30:01 crc kubenswrapper[4677]: I1203 14:30:01.939413 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" event={"ID":"ef734bef-9a2b-4622-b8e8-d476cebd82ee","Type":"ContainerDied","Data":"325ba1ba74b7328ef77ee4badd32e67dbe7042d8c02ab66efa27598016342674"} Dec 03 14:30:01 crc kubenswrapper[4677]: I1203 14:30:01.939535 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" event={"ID":"ef734bef-9a2b-4622-b8e8-d476cebd82ee","Type":"ContainerStarted","Data":"79a068afc69c4029f15c84f1941cfd146eefe5a5d1246060db922958a72bc037"} Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.313577 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.434849 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef734bef-9a2b-4622-b8e8-d476cebd82ee-config-volume\") pod \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.435004 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xggmb\" (UniqueName: \"kubernetes.io/projected/ef734bef-9a2b-4622-b8e8-d476cebd82ee-kube-api-access-xggmb\") pod \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.435093 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef734bef-9a2b-4622-b8e8-d476cebd82ee-secret-volume\") pod \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\" (UID: \"ef734bef-9a2b-4622-b8e8-d476cebd82ee\") " Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.436243 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef734bef-9a2b-4622-b8e8-d476cebd82ee-config-volume" (OuterVolumeSpecName: "config-volume") pod "ef734bef-9a2b-4622-b8e8-d476cebd82ee" (UID: "ef734bef-9a2b-4622-b8e8-d476cebd82ee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.442527 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef734bef-9a2b-4622-b8e8-d476cebd82ee-kube-api-access-xggmb" (OuterVolumeSpecName: "kube-api-access-xggmb") pod "ef734bef-9a2b-4622-b8e8-d476cebd82ee" (UID: "ef734bef-9a2b-4622-b8e8-d476cebd82ee"). InnerVolumeSpecName "kube-api-access-xggmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.456340 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef734bef-9a2b-4622-b8e8-d476cebd82ee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ef734bef-9a2b-4622-b8e8-d476cebd82ee" (UID: "ef734bef-9a2b-4622-b8e8-d476cebd82ee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.537310 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef734bef-9a2b-4622-b8e8-d476cebd82ee-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.537346 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xggmb\" (UniqueName: \"kubernetes.io/projected/ef734bef-9a2b-4622-b8e8-d476cebd82ee-kube-api-access-xggmb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.537358 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef734bef-9a2b-4622-b8e8-d476cebd82ee-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.963721 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" event={"ID":"ef734bef-9a2b-4622-b8e8-d476cebd82ee","Type":"ContainerDied","Data":"79a068afc69c4029f15c84f1941cfd146eefe5a5d1246060db922958a72bc037"} Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.963773 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms" Dec 03 14:30:03 crc kubenswrapper[4677]: I1203 14:30:03.963781 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79a068afc69c4029f15c84f1941cfd146eefe5a5d1246060db922958a72bc037" Dec 03 14:30:04 crc kubenswrapper[4677]: I1203 14:30:04.408779 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52"] Dec 03 14:30:04 crc kubenswrapper[4677]: I1203 14:30:04.417150 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-drx52"] Dec 03 14:30:05 crc kubenswrapper[4677]: I1203 14:30:05.991325 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e893d3c-7f15-44a6-b881-f658c0e06171" path="/var/lib/kubelet/pods/5e893d3c-7f15-44a6-b881-f658c0e06171/volumes" Dec 03 14:30:07 crc kubenswrapper[4677]: I1203 14:30:07.157603 4677 scope.go:117] "RemoveContainer" containerID="f7bdd33c43e30e88ef4fbe6b4ad11ae175d53bdb573475dbf86d0ae528cdd267" Dec 03 14:30:09 crc kubenswrapper[4677]: I1203 14:30:09.983908 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:30:09 crc kubenswrapper[4677]: E1203 14:30:09.984654 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:30:21 crc kubenswrapper[4677]: I1203 14:30:21.979048 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:30:21 crc kubenswrapper[4677]: E1203 14:30:21.979969 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:30:36 crc kubenswrapper[4677]: I1203 14:30:36.976643 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:30:36 crc kubenswrapper[4677]: E1203 14:30:36.977574 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:30:47 crc kubenswrapper[4677]: I1203 14:30:47.975626 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:30:49 crc kubenswrapper[4677]: I1203 14:30:49.382835 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"2d41f9dabdf4731fdfeb081c4a938a6ca90a305c55b57b82cac90a2c2d9cb5c5"} Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.055886 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-67cvl"] Dec 03 14:31:06 crc kubenswrapper[4677]: E1203 14:31:06.057331 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef734bef-9a2b-4622-b8e8-d476cebd82ee" containerName="collect-profiles" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.057348 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef734bef-9a2b-4622-b8e8-d476cebd82ee" containerName="collect-profiles" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.057623 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef734bef-9a2b-4622-b8e8-d476cebd82ee" containerName="collect-profiles" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.059597 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.083373 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-67cvl"] Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.125109 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-utilities\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.125443 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcp8x\" (UniqueName: \"kubernetes.io/projected/191817fc-3ce2-496a-a890-fdc846d81781-kube-api-access-fcp8x\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.125463 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-catalog-content\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.227150 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-utilities\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.227208 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcp8x\" (UniqueName: \"kubernetes.io/projected/191817fc-3ce2-496a-a890-fdc846d81781-kube-api-access-fcp8x\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.227237 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-catalog-content\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.227921 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-catalog-content\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.228227 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-utilities\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.254848 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcp8x\" (UniqueName: \"kubernetes.io/projected/191817fc-3ce2-496a-a890-fdc846d81781-kube-api-access-fcp8x\") pod \"redhat-marketplace-67cvl\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.382112 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:06 crc kubenswrapper[4677]: I1203 14:31:06.867213 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-67cvl"] Dec 03 14:31:07 crc kubenswrapper[4677]: I1203 14:31:07.562452 4677 generic.go:334] "Generic (PLEG): container finished" podID="191817fc-3ce2-496a-a890-fdc846d81781" containerID="5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12" exitCode=0 Dec 03 14:31:07 crc kubenswrapper[4677]: I1203 14:31:07.562504 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67cvl" event={"ID":"191817fc-3ce2-496a-a890-fdc846d81781","Type":"ContainerDied","Data":"5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12"} Dec 03 14:31:07 crc kubenswrapper[4677]: I1203 14:31:07.562542 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67cvl" event={"ID":"191817fc-3ce2-496a-a890-fdc846d81781","Type":"ContainerStarted","Data":"e8e6b7237f1b5e464f9dd2d705e54a4f74b94880fd4feb0697a97de89443cd1b"} Dec 03 14:31:07 crc kubenswrapper[4677]: I1203 14:31:07.564431 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:31:08 crc kubenswrapper[4677]: I1203 14:31:08.576740 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67cvl" event={"ID":"191817fc-3ce2-496a-a890-fdc846d81781","Type":"ContainerStarted","Data":"348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02"} Dec 03 14:31:09 crc kubenswrapper[4677]: I1203 14:31:09.588295 4677 generic.go:334] "Generic (PLEG): container finished" podID="191817fc-3ce2-496a-a890-fdc846d81781" containerID="348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02" exitCode=0 Dec 03 14:31:09 crc kubenswrapper[4677]: I1203 14:31:09.588354 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67cvl" event={"ID":"191817fc-3ce2-496a-a890-fdc846d81781","Type":"ContainerDied","Data":"348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02"} Dec 03 14:31:10 crc kubenswrapper[4677]: I1203 14:31:10.598693 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67cvl" event={"ID":"191817fc-3ce2-496a-a890-fdc846d81781","Type":"ContainerStarted","Data":"7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc"} Dec 03 14:31:10 crc kubenswrapper[4677]: I1203 14:31:10.630802 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-67cvl" podStartSLOduration=2.228268738 podStartE2EDuration="4.630784036s" podCreationTimestamp="2025-12-03 14:31:06 +0000 UTC" firstStartedPulling="2025-12-03 14:31:07.564034098 +0000 UTC m=+2658.310366553" lastFinishedPulling="2025-12-03 14:31:09.966549406 +0000 UTC m=+2660.712881851" observedRunningTime="2025-12-03 14:31:10.623405932 +0000 UTC m=+2661.369738407" watchObservedRunningTime="2025-12-03 14:31:10.630784036 +0000 UTC m=+2661.377116491" Dec 03 14:31:16 crc kubenswrapper[4677]: I1203 14:31:16.383176 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:16 crc kubenswrapper[4677]: I1203 14:31:16.383810 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:16 crc kubenswrapper[4677]: I1203 14:31:16.440182 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:16 crc kubenswrapper[4677]: I1203 14:31:16.726930 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:16 crc kubenswrapper[4677]: I1203 14:31:16.800676 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-67cvl"] Dec 03 14:31:18 crc kubenswrapper[4677]: I1203 14:31:18.693290 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-67cvl" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="registry-server" containerID="cri-o://7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc" gracePeriod=2 Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.121633 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.288659 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-utilities\") pod \"191817fc-3ce2-496a-a890-fdc846d81781\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.288720 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcp8x\" (UniqueName: \"kubernetes.io/projected/191817fc-3ce2-496a-a890-fdc846d81781-kube-api-access-fcp8x\") pod \"191817fc-3ce2-496a-a890-fdc846d81781\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.288736 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-catalog-content\") pod \"191817fc-3ce2-496a-a890-fdc846d81781\" (UID: \"191817fc-3ce2-496a-a890-fdc846d81781\") " Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.289894 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-utilities" (OuterVolumeSpecName: "utilities") pod "191817fc-3ce2-496a-a890-fdc846d81781" (UID: "191817fc-3ce2-496a-a890-fdc846d81781"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.298611 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/191817fc-3ce2-496a-a890-fdc846d81781-kube-api-access-fcp8x" (OuterVolumeSpecName: "kube-api-access-fcp8x") pod "191817fc-3ce2-496a-a890-fdc846d81781" (UID: "191817fc-3ce2-496a-a890-fdc846d81781"). InnerVolumeSpecName "kube-api-access-fcp8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.308154 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "191817fc-3ce2-496a-a890-fdc846d81781" (UID: "191817fc-3ce2-496a-a890-fdc846d81781"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.391225 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.391260 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcp8x\" (UniqueName: \"kubernetes.io/projected/191817fc-3ce2-496a-a890-fdc846d81781-kube-api-access-fcp8x\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.391275 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191817fc-3ce2-496a-a890-fdc846d81781-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.704992 4677 generic.go:334] "Generic (PLEG): container finished" podID="191817fc-3ce2-496a-a890-fdc846d81781" containerID="7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc" exitCode=0 Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.705033 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67cvl" event={"ID":"191817fc-3ce2-496a-a890-fdc846d81781","Type":"ContainerDied","Data":"7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc"} Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.705059 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67cvl" event={"ID":"191817fc-3ce2-496a-a890-fdc846d81781","Type":"ContainerDied","Data":"e8e6b7237f1b5e464f9dd2d705e54a4f74b94880fd4feb0697a97de89443cd1b"} Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.705098 4677 scope.go:117] "RemoveContainer" containerID="7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.705092 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67cvl" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.733188 4677 scope.go:117] "RemoveContainer" containerID="348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.741160 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-67cvl"] Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.757695 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-67cvl"] Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.768619 4677 scope.go:117] "RemoveContainer" containerID="5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.802622 4677 scope.go:117] "RemoveContainer" containerID="7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc" Dec 03 14:31:19 crc kubenswrapper[4677]: E1203 14:31:19.803367 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc\": container with ID starting with 7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc not found: ID does not exist" containerID="7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.803436 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc"} err="failed to get container status \"7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc\": rpc error: code = NotFound desc = could not find container \"7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc\": container with ID starting with 7fb03bf2a4aa0f7f1b7610a378407197592a937dd89a3cebee2523949827abdc not found: ID does not exist" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.803464 4677 scope.go:117] "RemoveContainer" containerID="348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02" Dec 03 14:31:19 crc kubenswrapper[4677]: E1203 14:31:19.803743 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02\": container with ID starting with 348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02 not found: ID does not exist" containerID="348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.803762 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02"} err="failed to get container status \"348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02\": rpc error: code = NotFound desc = could not find container \"348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02\": container with ID starting with 348f04859492c8e42c63a0ed57844e2878a6cf44e1fcd0643562b0168f286a02 not found: ID does not exist" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.803775 4677 scope.go:117] "RemoveContainer" containerID="5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12" Dec 03 14:31:19 crc kubenswrapper[4677]: E1203 14:31:19.803936 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12\": container with ID starting with 5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12 not found: ID does not exist" containerID="5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.803968 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12"} err="failed to get container status \"5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12\": rpc error: code = NotFound desc = could not find container \"5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12\": container with ID starting with 5b787f9d3cc50d578228fb48d3fff83393846f9d3bc91e12ce9e874dc164bc12 not found: ID does not exist" Dec 03 14:31:19 crc kubenswrapper[4677]: I1203 14:31:19.986229 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="191817fc-3ce2-496a-a890-fdc846d81781" path="/var/lib/kubelet/pods/191817fc-3ce2-496a-a890-fdc846d81781/volumes" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.200403 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2ld4f"] Dec 03 14:31:25 crc kubenswrapper[4677]: E1203 14:31:25.201505 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="registry-server" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.201531 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="registry-server" Dec 03 14:31:25 crc kubenswrapper[4677]: E1203 14:31:25.201568 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="extract-content" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.201592 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="extract-content" Dec 03 14:31:25 crc kubenswrapper[4677]: E1203 14:31:25.201621 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="extract-utilities" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.201634 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="extract-utilities" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.201966 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="191817fc-3ce2-496a-a890-fdc846d81781" containerName="registry-server" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.203882 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.210861 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2ld4f"] Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.306366 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-utilities\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.306538 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-catalog-content\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.306668 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nb5c\" (UniqueName: \"kubernetes.io/projected/4f644783-b94b-4563-b6be-4459338d9327-kube-api-access-7nb5c\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.408517 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-utilities\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.408605 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-catalog-content\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.408670 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nb5c\" (UniqueName: \"kubernetes.io/projected/4f644783-b94b-4563-b6be-4459338d9327-kube-api-access-7nb5c\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.409460 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-utilities\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.409563 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-catalog-content\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.429743 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nb5c\" (UniqueName: \"kubernetes.io/projected/4f644783-b94b-4563-b6be-4459338d9327-kube-api-access-7nb5c\") pod \"community-operators-2ld4f\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:25 crc kubenswrapper[4677]: I1203 14:31:25.529358 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:26 crc kubenswrapper[4677]: I1203 14:31:26.140591 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2ld4f"] Dec 03 14:31:26 crc kubenswrapper[4677]: I1203 14:31:26.779613 4677 generic.go:334] "Generic (PLEG): container finished" podID="4f644783-b94b-4563-b6be-4459338d9327" containerID="09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176" exitCode=0 Dec 03 14:31:26 crc kubenswrapper[4677]: I1203 14:31:26.780163 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ld4f" event={"ID":"4f644783-b94b-4563-b6be-4459338d9327","Type":"ContainerDied","Data":"09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176"} Dec 03 14:31:26 crc kubenswrapper[4677]: I1203 14:31:26.780196 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ld4f" event={"ID":"4f644783-b94b-4563-b6be-4459338d9327","Type":"ContainerStarted","Data":"cb87dbaa8221d74177a24125468d2efb03ea664021d31561dbf226a71c9288f4"} Dec 03 14:31:27 crc kubenswrapper[4677]: I1203 14:31:27.791901 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ld4f" event={"ID":"4f644783-b94b-4563-b6be-4459338d9327","Type":"ContainerStarted","Data":"ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939"} Dec 03 14:31:28 crc kubenswrapper[4677]: I1203 14:31:28.804563 4677 generic.go:334] "Generic (PLEG): container finished" podID="4f644783-b94b-4563-b6be-4459338d9327" containerID="ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939" exitCode=0 Dec 03 14:31:28 crc kubenswrapper[4677]: I1203 14:31:28.804626 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ld4f" event={"ID":"4f644783-b94b-4563-b6be-4459338d9327","Type":"ContainerDied","Data":"ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939"} Dec 03 14:31:29 crc kubenswrapper[4677]: I1203 14:31:29.814083 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ld4f" event={"ID":"4f644783-b94b-4563-b6be-4459338d9327","Type":"ContainerStarted","Data":"3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed"} Dec 03 14:31:29 crc kubenswrapper[4677]: I1203 14:31:29.841009 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2ld4f" podStartSLOduration=2.402057368 podStartE2EDuration="4.84099125s" podCreationTimestamp="2025-12-03 14:31:25 +0000 UTC" firstStartedPulling="2025-12-03 14:31:26.782668086 +0000 UTC m=+2677.529000541" lastFinishedPulling="2025-12-03 14:31:29.221601968 +0000 UTC m=+2679.967934423" observedRunningTime="2025-12-03 14:31:29.833396641 +0000 UTC m=+2680.579729096" watchObservedRunningTime="2025-12-03 14:31:29.84099125 +0000 UTC m=+2680.587323715" Dec 03 14:31:35 crc kubenswrapper[4677]: I1203 14:31:35.530240 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:35 crc kubenswrapper[4677]: I1203 14:31:35.530830 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:35 crc kubenswrapper[4677]: I1203 14:31:35.580623 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:35 crc kubenswrapper[4677]: I1203 14:31:35.927630 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:35 crc kubenswrapper[4677]: I1203 14:31:35.988160 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2ld4f"] Dec 03 14:31:37 crc kubenswrapper[4677]: I1203 14:31:37.890864 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2ld4f" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="registry-server" containerID="cri-o://3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed" gracePeriod=2 Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.391864 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.573660 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-catalog-content\") pod \"4f644783-b94b-4563-b6be-4459338d9327\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.573868 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-utilities\") pod \"4f644783-b94b-4563-b6be-4459338d9327\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.573963 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nb5c\" (UniqueName: \"kubernetes.io/projected/4f644783-b94b-4563-b6be-4459338d9327-kube-api-access-7nb5c\") pod \"4f644783-b94b-4563-b6be-4459338d9327\" (UID: \"4f644783-b94b-4563-b6be-4459338d9327\") " Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.575142 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-utilities" (OuterVolumeSpecName: "utilities") pod "4f644783-b94b-4563-b6be-4459338d9327" (UID: "4f644783-b94b-4563-b6be-4459338d9327"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.576207 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.581245 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f644783-b94b-4563-b6be-4459338d9327-kube-api-access-7nb5c" (OuterVolumeSpecName: "kube-api-access-7nb5c") pod "4f644783-b94b-4563-b6be-4459338d9327" (UID: "4f644783-b94b-4563-b6be-4459338d9327"). InnerVolumeSpecName "kube-api-access-7nb5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.631232 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f644783-b94b-4563-b6be-4459338d9327" (UID: "4f644783-b94b-4563-b6be-4459338d9327"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.678054 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f644783-b94b-4563-b6be-4459338d9327-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.678107 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nb5c\" (UniqueName: \"kubernetes.io/projected/4f644783-b94b-4563-b6be-4459338d9327-kube-api-access-7nb5c\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.901011 4677 generic.go:334] "Generic (PLEG): container finished" podID="4f644783-b94b-4563-b6be-4459338d9327" containerID="3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed" exitCode=0 Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.901062 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ld4f" event={"ID":"4f644783-b94b-4563-b6be-4459338d9327","Type":"ContainerDied","Data":"3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed"} Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.901066 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ld4f" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.901088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ld4f" event={"ID":"4f644783-b94b-4563-b6be-4459338d9327","Type":"ContainerDied","Data":"cb87dbaa8221d74177a24125468d2efb03ea664021d31561dbf226a71c9288f4"} Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.901108 4677 scope.go:117] "RemoveContainer" containerID="3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.930896 4677 scope.go:117] "RemoveContainer" containerID="ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.942341 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2ld4f"] Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.974164 4677 scope.go:117] "RemoveContainer" containerID="09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176" Dec 03 14:31:38 crc kubenswrapper[4677]: I1203 14:31:38.986475 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2ld4f"] Dec 03 14:31:39 crc kubenswrapper[4677]: I1203 14:31:39.016041 4677 scope.go:117] "RemoveContainer" containerID="3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed" Dec 03 14:31:39 crc kubenswrapper[4677]: E1203 14:31:39.016852 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed\": container with ID starting with 3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed not found: ID does not exist" containerID="3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed" Dec 03 14:31:39 crc kubenswrapper[4677]: I1203 14:31:39.016935 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed"} err="failed to get container status \"3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed\": rpc error: code = NotFound desc = could not find container \"3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed\": container with ID starting with 3cb15e648f0f50c0f7fecdeb4104e32ea88d6263e88cc7ffad32b0bef2b5cbed not found: ID does not exist" Dec 03 14:31:39 crc kubenswrapper[4677]: I1203 14:31:39.017007 4677 scope.go:117] "RemoveContainer" containerID="ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939" Dec 03 14:31:39 crc kubenswrapper[4677]: E1203 14:31:39.017612 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939\": container with ID starting with ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939 not found: ID does not exist" containerID="ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939" Dec 03 14:31:39 crc kubenswrapper[4677]: I1203 14:31:39.017663 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939"} err="failed to get container status \"ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939\": rpc error: code = NotFound desc = could not find container \"ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939\": container with ID starting with ef6e9ee5a7e37cd5f270786871c0f56c29ff41c12501f79dc92111f082ad8939 not found: ID does not exist" Dec 03 14:31:39 crc kubenswrapper[4677]: I1203 14:31:39.017698 4677 scope.go:117] "RemoveContainer" containerID="09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176" Dec 03 14:31:39 crc kubenswrapper[4677]: E1203 14:31:39.018073 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176\": container with ID starting with 09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176 not found: ID does not exist" containerID="09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176" Dec 03 14:31:39 crc kubenswrapper[4677]: I1203 14:31:39.018112 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176"} err="failed to get container status \"09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176\": rpc error: code = NotFound desc = could not find container \"09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176\": container with ID starting with 09c4b72de40a0226b7d9ca91508a84669d1e2b598a5d800006b1401c6487f176 not found: ID does not exist" Dec 03 14:31:39 crc kubenswrapper[4677]: I1203 14:31:39.988598 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f644783-b94b-4563-b6be-4459338d9327" path="/var/lib/kubelet/pods/4f644783-b94b-4563-b6be-4459338d9327/volumes" Dec 03 14:33:08 crc kubenswrapper[4677]: I1203 14:33:08.437106 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:33:08 crc kubenswrapper[4677]: I1203 14:33:08.437722 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:33:32 crc kubenswrapper[4677]: I1203 14:33:32.966366 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nvd4x"] Dec 03 14:33:32 crc kubenswrapper[4677]: E1203 14:33:32.967460 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="registry-server" Dec 03 14:33:32 crc kubenswrapper[4677]: I1203 14:33:32.967476 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="registry-server" Dec 03 14:33:32 crc kubenswrapper[4677]: E1203 14:33:32.967489 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="extract-content" Dec 03 14:33:32 crc kubenswrapper[4677]: I1203 14:33:32.967495 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="extract-content" Dec 03 14:33:32 crc kubenswrapper[4677]: E1203 14:33:32.967508 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="extract-utilities" Dec 03 14:33:32 crc kubenswrapper[4677]: I1203 14:33:32.967516 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="extract-utilities" Dec 03 14:33:32 crc kubenswrapper[4677]: I1203 14:33:32.967766 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f644783-b94b-4563-b6be-4459338d9327" containerName="registry-server" Dec 03 14:33:32 crc kubenswrapper[4677]: I1203 14:33:32.969511 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:32 crc kubenswrapper[4677]: I1203 14:33:32.980616 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nvd4x"] Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.133545 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mns\" (UniqueName: \"kubernetes.io/projected/88446b28-8a27-4ae2-b6a7-e054cc5a6175-kube-api-access-v7mns\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.133739 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-utilities\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.133876 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-catalog-content\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.236311 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-utilities\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.236376 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-catalog-content\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.236539 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mns\" (UniqueName: \"kubernetes.io/projected/88446b28-8a27-4ae2-b6a7-e054cc5a6175-kube-api-access-v7mns\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.237248 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-catalog-content\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.238064 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-utilities\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.262470 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mns\" (UniqueName: \"kubernetes.io/projected/88446b28-8a27-4ae2-b6a7-e054cc5a6175-kube-api-access-v7mns\") pod \"redhat-operators-nvd4x\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.289865 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:33 crc kubenswrapper[4677]: I1203 14:33:33.753424 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nvd4x"] Dec 03 14:33:34 crc kubenswrapper[4677]: I1203 14:33:34.061185 4677 generic.go:334] "Generic (PLEG): container finished" podID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerID="f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5" exitCode=0 Dec 03 14:33:34 crc kubenswrapper[4677]: I1203 14:33:34.061332 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvd4x" event={"ID":"88446b28-8a27-4ae2-b6a7-e054cc5a6175","Type":"ContainerDied","Data":"f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5"} Dec 03 14:33:34 crc kubenswrapper[4677]: I1203 14:33:34.061499 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvd4x" event={"ID":"88446b28-8a27-4ae2-b6a7-e054cc5a6175","Type":"ContainerStarted","Data":"67aaeacffdffefae6ab0039fec87b4a6063afb20fbbac08ee869d18b1d05525e"} Dec 03 14:33:35 crc kubenswrapper[4677]: I1203 14:33:35.074024 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvd4x" event={"ID":"88446b28-8a27-4ae2-b6a7-e054cc5a6175","Type":"ContainerStarted","Data":"2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f"} Dec 03 14:33:37 crc kubenswrapper[4677]: I1203 14:33:37.092628 4677 generic.go:334] "Generic (PLEG): container finished" podID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerID="2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f" exitCode=0 Dec 03 14:33:37 crc kubenswrapper[4677]: I1203 14:33:37.092687 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvd4x" event={"ID":"88446b28-8a27-4ae2-b6a7-e054cc5a6175","Type":"ContainerDied","Data":"2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f"} Dec 03 14:33:38 crc kubenswrapper[4677]: I1203 14:33:38.437766 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:33:38 crc kubenswrapper[4677]: I1203 14:33:38.438283 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:33:40 crc kubenswrapper[4677]: I1203 14:33:40.123464 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvd4x" event={"ID":"88446b28-8a27-4ae2-b6a7-e054cc5a6175","Type":"ContainerStarted","Data":"35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080"} Dec 03 14:33:40 crc kubenswrapper[4677]: I1203 14:33:40.147620 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nvd4x" podStartSLOduration=3.177910886 podStartE2EDuration="8.147602911s" podCreationTimestamp="2025-12-03 14:33:32 +0000 UTC" firstStartedPulling="2025-12-03 14:33:34.06773633 +0000 UTC m=+2804.814068795" lastFinishedPulling="2025-12-03 14:33:39.037428365 +0000 UTC m=+2809.783760820" observedRunningTime="2025-12-03 14:33:40.142001987 +0000 UTC m=+2810.888334442" watchObservedRunningTime="2025-12-03 14:33:40.147602911 +0000 UTC m=+2810.893935366" Dec 03 14:33:43 crc kubenswrapper[4677]: I1203 14:33:43.290640 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:43 crc kubenswrapper[4677]: I1203 14:33:43.292039 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:44 crc kubenswrapper[4677]: I1203 14:33:44.345105 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nvd4x" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="registry-server" probeResult="failure" output=< Dec 03 14:33:44 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 14:33:44 crc kubenswrapper[4677]: > Dec 03 14:33:44 crc kubenswrapper[4677]: E1203 14:33:44.583640 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod891909bd_1955_436b_8664_79a41b0676e5.slice/crio-conmon-d0941009e801c12ff036b3576e2260a81f782c1e944a4fb3db327474159378af.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod891909bd_1955_436b_8664_79a41b0676e5.slice/crio-d0941009e801c12ff036b3576e2260a81f782c1e944a4fb3db327474159378af.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:33:45 crc kubenswrapper[4677]: I1203 14:33:45.248571 4677 generic.go:334] "Generic (PLEG): container finished" podID="891909bd-1955-436b-8664-79a41b0676e5" containerID="d0941009e801c12ff036b3576e2260a81f782c1e944a4fb3db327474159378af" exitCode=0 Dec 03 14:33:45 crc kubenswrapper[4677]: I1203 14:33:45.248663 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" event={"ID":"891909bd-1955-436b-8664-79a41b0676e5","Type":"ContainerDied","Data":"d0941009e801c12ff036b3576e2260a81f782c1e944a4fb3db327474159378af"} Dec 03 14:33:46 crc kubenswrapper[4677]: I1203 14:33:46.873387 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.029102 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2259\" (UniqueName: \"kubernetes.io/projected/891909bd-1955-436b-8664-79a41b0676e5-kube-api-access-z2259\") pod \"891909bd-1955-436b-8664-79a41b0676e5\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.029498 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-secret-0\") pod \"891909bd-1955-436b-8664-79a41b0676e5\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.029598 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-inventory\") pod \"891909bd-1955-436b-8664-79a41b0676e5\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.029657 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-ssh-key\") pod \"891909bd-1955-436b-8664-79a41b0676e5\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.029716 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-combined-ca-bundle\") pod \"891909bd-1955-436b-8664-79a41b0676e5\" (UID: \"891909bd-1955-436b-8664-79a41b0676e5\") " Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.049062 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "891909bd-1955-436b-8664-79a41b0676e5" (UID: "891909bd-1955-436b-8664-79a41b0676e5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.050103 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/891909bd-1955-436b-8664-79a41b0676e5-kube-api-access-z2259" (OuterVolumeSpecName: "kube-api-access-z2259") pod "891909bd-1955-436b-8664-79a41b0676e5" (UID: "891909bd-1955-436b-8664-79a41b0676e5"). InnerVolumeSpecName "kube-api-access-z2259". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.062247 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "891909bd-1955-436b-8664-79a41b0676e5" (UID: "891909bd-1955-436b-8664-79a41b0676e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.066472 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-inventory" (OuterVolumeSpecName: "inventory") pod "891909bd-1955-436b-8664-79a41b0676e5" (UID: "891909bd-1955-436b-8664-79a41b0676e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.069156 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "891909bd-1955-436b-8664-79a41b0676e5" (UID: "891909bd-1955-436b-8664-79a41b0676e5"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.131857 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.131896 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.131909 4677 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.131924 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2259\" (UniqueName: \"kubernetes.io/projected/891909bd-1955-436b-8664-79a41b0676e5-kube-api-access-z2259\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.131938 4677 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/891909bd-1955-436b-8664-79a41b0676e5-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.271332 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" event={"ID":"891909bd-1955-436b-8664-79a41b0676e5","Type":"ContainerDied","Data":"ed362823d683e1b09b976c0272a8662b9c3e9b1238383114799333cf2167f0d7"} Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.271386 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed362823d683e1b09b976c0272a8662b9c3e9b1238383114799333cf2167f0d7" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.271398 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-64ds6" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.370085 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl"] Dec 03 14:33:47 crc kubenswrapper[4677]: E1203 14:33:47.370519 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="891909bd-1955-436b-8664-79a41b0676e5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.370541 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="891909bd-1955-436b-8664-79a41b0676e5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.370786 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="891909bd-1955-436b-8664-79a41b0676e5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.371565 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.374464 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.378582 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.378942 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.378609 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.379313 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.379448 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.379689 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.390637 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl"] Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437319 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437483 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rhfx\" (UniqueName: \"kubernetes.io/projected/7d24039e-bd02-41b9-86b3-1b23b50f6824-kube-api-access-7rhfx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437635 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437710 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437736 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437769 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437833 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.437871 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.438050 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.539712 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.539800 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.539829 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.539861 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.539907 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.539961 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.540033 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.540090 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.540145 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rhfx\" (UniqueName: \"kubernetes.io/projected/7d24039e-bd02-41b9-86b3-1b23b50f6824-kube-api-access-7rhfx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.542000 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.545403 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.545655 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.545673 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.545803 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.546751 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.547045 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.547684 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.559411 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rhfx\" (UniqueName: \"kubernetes.io/projected/7d24039e-bd02-41b9-86b3-1b23b50f6824-kube-api-access-7rhfx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hcrwl\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:47 crc kubenswrapper[4677]: I1203 14:33:47.696657 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:33:48 crc kubenswrapper[4677]: I1203 14:33:48.240784 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl"] Dec 03 14:33:48 crc kubenswrapper[4677]: I1203 14:33:48.281361 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" event={"ID":"7d24039e-bd02-41b9-86b3-1b23b50f6824","Type":"ContainerStarted","Data":"294a7cc4102e6db1d0c097128920bd5692a3111f9a5f096ad761d59ba5621417"} Dec 03 14:33:50 crc kubenswrapper[4677]: I1203 14:33:50.312780 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" event={"ID":"7d24039e-bd02-41b9-86b3-1b23b50f6824","Type":"ContainerStarted","Data":"536a3a6b6b8f7edb5e3dec25c4faa52a329439e0589527fb8a649182b656c087"} Dec 03 14:33:50 crc kubenswrapper[4677]: I1203 14:33:50.338468 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" podStartSLOduration=2.455312039 podStartE2EDuration="3.338448571s" podCreationTimestamp="2025-12-03 14:33:47 +0000 UTC" firstStartedPulling="2025-12-03 14:33:48.245868872 +0000 UTC m=+2818.992201327" lastFinishedPulling="2025-12-03 14:33:49.129005404 +0000 UTC m=+2819.875337859" observedRunningTime="2025-12-03 14:33:50.328839797 +0000 UTC m=+2821.075172262" watchObservedRunningTime="2025-12-03 14:33:50.338448571 +0000 UTC m=+2821.084781036" Dec 03 14:33:53 crc kubenswrapper[4677]: I1203 14:33:53.341576 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:53 crc kubenswrapper[4677]: I1203 14:33:53.394579 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:53 crc kubenswrapper[4677]: I1203 14:33:53.578030 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nvd4x"] Dec 03 14:33:55 crc kubenswrapper[4677]: I1203 14:33:55.356136 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nvd4x" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="registry-server" containerID="cri-o://35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080" gracePeriod=2 Dec 03 14:33:55 crc kubenswrapper[4677]: I1203 14:33:55.839578 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:55 crc kubenswrapper[4677]: I1203 14:33:55.906154 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-catalog-content\") pod \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " Dec 03 14:33:55 crc kubenswrapper[4677]: I1203 14:33:55.906258 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-utilities\") pod \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " Dec 03 14:33:55 crc kubenswrapper[4677]: I1203 14:33:55.906406 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7mns\" (UniqueName: \"kubernetes.io/projected/88446b28-8a27-4ae2-b6a7-e054cc5a6175-kube-api-access-v7mns\") pod \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\" (UID: \"88446b28-8a27-4ae2-b6a7-e054cc5a6175\") " Dec 03 14:33:55 crc kubenswrapper[4677]: I1203 14:33:55.907224 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-utilities" (OuterVolumeSpecName: "utilities") pod "88446b28-8a27-4ae2-b6a7-e054cc5a6175" (UID: "88446b28-8a27-4ae2-b6a7-e054cc5a6175"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:55 crc kubenswrapper[4677]: I1203 14:33:55.915121 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88446b28-8a27-4ae2-b6a7-e054cc5a6175-kube-api-access-v7mns" (OuterVolumeSpecName: "kube-api-access-v7mns") pod "88446b28-8a27-4ae2-b6a7-e054cc5a6175" (UID: "88446b28-8a27-4ae2-b6a7-e054cc5a6175"). InnerVolumeSpecName "kube-api-access-v7mns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.011768 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.011809 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7mns\" (UniqueName: \"kubernetes.io/projected/88446b28-8a27-4ae2-b6a7-e054cc5a6175-kube-api-access-v7mns\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.061386 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88446b28-8a27-4ae2-b6a7-e054cc5a6175" (UID: "88446b28-8a27-4ae2-b6a7-e054cc5a6175"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.114080 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88446b28-8a27-4ae2-b6a7-e054cc5a6175-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.370920 4677 generic.go:334] "Generic (PLEG): container finished" podID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerID="35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080" exitCode=0 Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.371019 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvd4x" event={"ID":"88446b28-8a27-4ae2-b6a7-e054cc5a6175","Type":"ContainerDied","Data":"35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080"} Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.371068 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvd4x" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.371119 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvd4x" event={"ID":"88446b28-8a27-4ae2-b6a7-e054cc5a6175","Type":"ContainerDied","Data":"67aaeacffdffefae6ab0039fec87b4a6063afb20fbbac08ee869d18b1d05525e"} Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.371155 4677 scope.go:117] "RemoveContainer" containerID="35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.395312 4677 scope.go:117] "RemoveContainer" containerID="2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.415168 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nvd4x"] Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.429468 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nvd4x"] Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.433302 4677 scope.go:117] "RemoveContainer" containerID="f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.474602 4677 scope.go:117] "RemoveContainer" containerID="35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080" Dec 03 14:33:56 crc kubenswrapper[4677]: E1203 14:33:56.475152 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080\": container with ID starting with 35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080 not found: ID does not exist" containerID="35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.475190 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080"} err="failed to get container status \"35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080\": rpc error: code = NotFound desc = could not find container \"35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080\": container with ID starting with 35e6f009277b745b384e735548768f135e970d1acf6fc3f15a0bd6507b764080 not found: ID does not exist" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.475211 4677 scope.go:117] "RemoveContainer" containerID="2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f" Dec 03 14:33:56 crc kubenswrapper[4677]: E1203 14:33:56.475589 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f\": container with ID starting with 2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f not found: ID does not exist" containerID="2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.475637 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f"} err="failed to get container status \"2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f\": rpc error: code = NotFound desc = could not find container \"2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f\": container with ID starting with 2b3cd50636cb93e8200df54d4ab7fe99696550c7a8239e7bc8fdfee4ed535a6f not found: ID does not exist" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.475671 4677 scope.go:117] "RemoveContainer" containerID="f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5" Dec 03 14:33:56 crc kubenswrapper[4677]: E1203 14:33:56.475979 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5\": container with ID starting with f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5 not found: ID does not exist" containerID="f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5" Dec 03 14:33:56 crc kubenswrapper[4677]: I1203 14:33:56.476008 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5"} err="failed to get container status \"f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5\": rpc error: code = NotFound desc = could not find container \"f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5\": container with ID starting with f336119e5b812f35663f7103fd33cff42d7bf8fea52cb592c858f422cebfc0c5 not found: ID does not exist" Dec 03 14:33:57 crc kubenswrapper[4677]: I1203 14:33:57.989570 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" path="/var/lib/kubelet/pods/88446b28-8a27-4ae2-b6a7-e054cc5a6175/volumes" Dec 03 14:34:08 crc kubenswrapper[4677]: I1203 14:34:08.438101 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:34:08 crc kubenswrapper[4677]: I1203 14:34:08.439245 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:34:08 crc kubenswrapper[4677]: I1203 14:34:08.439338 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:34:08 crc kubenswrapper[4677]: I1203 14:34:08.440719 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d41f9dabdf4731fdfeb081c4a938a6ca90a305c55b57b82cac90a2c2d9cb5c5"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:34:08 crc kubenswrapper[4677]: I1203 14:34:08.440804 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://2d41f9dabdf4731fdfeb081c4a938a6ca90a305c55b57b82cac90a2c2d9cb5c5" gracePeriod=600 Dec 03 14:34:09 crc kubenswrapper[4677]: I1203 14:34:09.508844 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="2d41f9dabdf4731fdfeb081c4a938a6ca90a305c55b57b82cac90a2c2d9cb5c5" exitCode=0 Dec 03 14:34:09 crc kubenswrapper[4677]: I1203 14:34:09.508934 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"2d41f9dabdf4731fdfeb081c4a938a6ca90a305c55b57b82cac90a2c2d9cb5c5"} Dec 03 14:34:09 crc kubenswrapper[4677]: I1203 14:34:09.509549 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4"} Dec 03 14:34:09 crc kubenswrapper[4677]: I1203 14:34:09.509572 4677 scope.go:117] "RemoveContainer" containerID="1e447a7c3d50a7a1f180fef8dd117a13f5a388ef4b929157c961b97463e6f62e" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.197475 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-llfls"] Dec 03 14:34:52 crc kubenswrapper[4677]: E1203 14:34:52.198508 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="extract-utilities" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.198530 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="extract-utilities" Dec 03 14:34:52 crc kubenswrapper[4677]: E1203 14:34:52.198545 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="extract-content" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.198554 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="extract-content" Dec 03 14:34:52 crc kubenswrapper[4677]: E1203 14:34:52.198581 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="registry-server" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.198589 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="registry-server" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.198787 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="88446b28-8a27-4ae2-b6a7-e054cc5a6175" containerName="registry-server" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.200471 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.214012 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llfls"] Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.312479 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5fm2\" (UniqueName: \"kubernetes.io/projected/04155b7a-f655-4013-a6d2-a06c9a3b3d80-kube-api-access-f5fm2\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.312544 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-utilities\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.312703 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-catalog-content\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.414482 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-catalog-content\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.414899 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5fm2\" (UniqueName: \"kubernetes.io/projected/04155b7a-f655-4013-a6d2-a06c9a3b3d80-kube-api-access-f5fm2\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.415047 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-utilities\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.415050 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-catalog-content\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.415478 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-utilities\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.437682 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5fm2\" (UniqueName: \"kubernetes.io/projected/04155b7a-f655-4013-a6d2-a06c9a3b3d80-kube-api-access-f5fm2\") pod \"certified-operators-llfls\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:52 crc kubenswrapper[4677]: I1203 14:34:52.524566 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:34:53 crc kubenswrapper[4677]: I1203 14:34:53.094899 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llfls"] Dec 03 14:34:53 crc kubenswrapper[4677]: I1203 14:34:53.158593 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llfls" event={"ID":"04155b7a-f655-4013-a6d2-a06c9a3b3d80","Type":"ContainerStarted","Data":"0b759f767b5e901f7cd9ba51ea50c0c2b69be80e59d8a8ef0f03b0b2783f9f8c"} Dec 03 14:34:54 crc kubenswrapper[4677]: I1203 14:34:54.178200 4677 generic.go:334] "Generic (PLEG): container finished" podID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerID="dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b" exitCode=0 Dec 03 14:34:54 crc kubenswrapper[4677]: I1203 14:34:54.178250 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llfls" event={"ID":"04155b7a-f655-4013-a6d2-a06c9a3b3d80","Type":"ContainerDied","Data":"dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b"} Dec 03 14:34:56 crc kubenswrapper[4677]: I1203 14:34:56.199890 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llfls" event={"ID":"04155b7a-f655-4013-a6d2-a06c9a3b3d80","Type":"ContainerStarted","Data":"bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5"} Dec 03 14:34:57 crc kubenswrapper[4677]: I1203 14:34:57.211047 4677 generic.go:334] "Generic (PLEG): container finished" podID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerID="bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5" exitCode=0 Dec 03 14:34:57 crc kubenswrapper[4677]: I1203 14:34:57.211109 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llfls" event={"ID":"04155b7a-f655-4013-a6d2-a06c9a3b3d80","Type":"ContainerDied","Data":"bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5"} Dec 03 14:34:58 crc kubenswrapper[4677]: I1203 14:34:58.222892 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llfls" event={"ID":"04155b7a-f655-4013-a6d2-a06c9a3b3d80","Type":"ContainerStarted","Data":"136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370"} Dec 03 14:34:58 crc kubenswrapper[4677]: I1203 14:34:58.249016 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-llfls" podStartSLOduration=2.818827729 podStartE2EDuration="6.248996378s" podCreationTimestamp="2025-12-03 14:34:52 +0000 UTC" firstStartedPulling="2025-12-03 14:34:54.179887704 +0000 UTC m=+2884.926220159" lastFinishedPulling="2025-12-03 14:34:57.610056353 +0000 UTC m=+2888.356388808" observedRunningTime="2025-12-03 14:34:58.238082968 +0000 UTC m=+2888.984415443" watchObservedRunningTime="2025-12-03 14:34:58.248996378 +0000 UTC m=+2888.995328843" Dec 03 14:35:02 crc kubenswrapper[4677]: I1203 14:35:02.524897 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:35:02 crc kubenswrapper[4677]: I1203 14:35:02.526532 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:35:02 crc kubenswrapper[4677]: I1203 14:35:02.582017 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:35:03 crc kubenswrapper[4677]: I1203 14:35:03.314611 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:35:03 crc kubenswrapper[4677]: I1203 14:35:03.367634 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llfls"] Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.285758 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-llfls" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="registry-server" containerID="cri-o://136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370" gracePeriod=2 Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.738263 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.893327 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5fm2\" (UniqueName: \"kubernetes.io/projected/04155b7a-f655-4013-a6d2-a06c9a3b3d80-kube-api-access-f5fm2\") pod \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.893618 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-utilities\") pod \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.893644 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-catalog-content\") pod \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\" (UID: \"04155b7a-f655-4013-a6d2-a06c9a3b3d80\") " Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.894598 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-utilities" (OuterVolumeSpecName: "utilities") pod "04155b7a-f655-4013-a6d2-a06c9a3b3d80" (UID: "04155b7a-f655-4013-a6d2-a06c9a3b3d80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.912619 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04155b7a-f655-4013-a6d2-a06c9a3b3d80-kube-api-access-f5fm2" (OuterVolumeSpecName: "kube-api-access-f5fm2") pod "04155b7a-f655-4013-a6d2-a06c9a3b3d80" (UID: "04155b7a-f655-4013-a6d2-a06c9a3b3d80"). InnerVolumeSpecName "kube-api-access-f5fm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.917771 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.917807 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5fm2\" (UniqueName: \"kubernetes.io/projected/04155b7a-f655-4013-a6d2-a06c9a3b3d80-kube-api-access-f5fm2\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:05 crc kubenswrapper[4677]: I1203 14:35:05.966220 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04155b7a-f655-4013-a6d2-a06c9a3b3d80" (UID: "04155b7a-f655-4013-a6d2-a06c9a3b3d80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.020244 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04155b7a-f655-4013-a6d2-a06c9a3b3d80-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.295683 4677 generic.go:334] "Generic (PLEG): container finished" podID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerID="136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370" exitCode=0 Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.295733 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llfls" event={"ID":"04155b7a-f655-4013-a6d2-a06c9a3b3d80","Type":"ContainerDied","Data":"136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370"} Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.295751 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llfls" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.295776 4677 scope.go:117] "RemoveContainer" containerID="136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.295763 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llfls" event={"ID":"04155b7a-f655-4013-a6d2-a06c9a3b3d80","Type":"ContainerDied","Data":"0b759f767b5e901f7cd9ba51ea50c0c2b69be80e59d8a8ef0f03b0b2783f9f8c"} Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.323044 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llfls"] Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.328535 4677 scope.go:117] "RemoveContainer" containerID="bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.336318 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-llfls"] Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.349741 4677 scope.go:117] "RemoveContainer" containerID="dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.395810 4677 scope.go:117] "RemoveContainer" containerID="136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370" Dec 03 14:35:06 crc kubenswrapper[4677]: E1203 14:35:06.396617 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370\": container with ID starting with 136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370 not found: ID does not exist" containerID="136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.396645 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370"} err="failed to get container status \"136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370\": rpc error: code = NotFound desc = could not find container \"136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370\": container with ID starting with 136c584625372da1031641e83c05f2df71d83b083260f77d5e07ab65b24c6370 not found: ID does not exist" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.396667 4677 scope.go:117] "RemoveContainer" containerID="bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5" Dec 03 14:35:06 crc kubenswrapper[4677]: E1203 14:35:06.396885 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5\": container with ID starting with bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5 not found: ID does not exist" containerID="bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.397011 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5"} err="failed to get container status \"bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5\": rpc error: code = NotFound desc = could not find container \"bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5\": container with ID starting with bb23e91b1d871c6d1df6ee5f64f9a0b4cda67aec9d5cb56ad5c44bd9b5a7dce5 not found: ID does not exist" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.397124 4677 scope.go:117] "RemoveContainer" containerID="dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b" Dec 03 14:35:06 crc kubenswrapper[4677]: E1203 14:35:06.397459 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b\": container with ID starting with dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b not found: ID does not exist" containerID="dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b" Dec 03 14:35:06 crc kubenswrapper[4677]: I1203 14:35:06.397481 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b"} err="failed to get container status \"dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b\": rpc error: code = NotFound desc = could not find container \"dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b\": container with ID starting with dee5c889511b549aa43af628c55ed8f3511b551f48c0587455084652680b028b not found: ID does not exist" Dec 03 14:35:07 crc kubenswrapper[4677]: I1203 14:35:07.988845 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" path="/var/lib/kubelet/pods/04155b7a-f655-4013-a6d2-a06c9a3b3d80/volumes" Dec 03 14:36:08 crc kubenswrapper[4677]: I1203 14:36:08.436810 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:36:08 crc kubenswrapper[4677]: I1203 14:36:08.437404 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:36:38 crc kubenswrapper[4677]: I1203 14:36:38.437740 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:36:38 crc kubenswrapper[4677]: I1203 14:36:38.438416 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:36:50 crc kubenswrapper[4677]: I1203 14:36:50.329764 4677 generic.go:334] "Generic (PLEG): container finished" podID="7d24039e-bd02-41b9-86b3-1b23b50f6824" containerID="536a3a6b6b8f7edb5e3dec25c4faa52a329439e0589527fb8a649182b656c087" exitCode=0 Dec 03 14:36:50 crc kubenswrapper[4677]: I1203 14:36:50.329900 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" event={"ID":"7d24039e-bd02-41b9-86b3-1b23b50f6824","Type":"ContainerDied","Data":"536a3a6b6b8f7edb5e3dec25c4faa52a329439e0589527fb8a649182b656c087"} Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.767624 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.812399 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-0\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.812561 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-extra-config-0\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.812588 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-inventory\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.812633 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-1\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.812655 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-combined-ca-bundle\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.819171 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.843291 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.846058 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.848646 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.850273 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-inventory" (OuterVolumeSpecName: "inventory") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915008 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-0\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915060 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rhfx\" (UniqueName: \"kubernetes.io/projected/7d24039e-bd02-41b9-86b3-1b23b50f6824-kube-api-access-7rhfx\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915080 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-1\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915129 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-ssh-key\") pod \"7d24039e-bd02-41b9-86b3-1b23b50f6824\" (UID: \"7d24039e-bd02-41b9-86b3-1b23b50f6824\") " Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915547 4677 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915561 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915572 4677 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915587 4677 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.915600 4677 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.921995 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d24039e-bd02-41b9-86b3-1b23b50f6824-kube-api-access-7rhfx" (OuterVolumeSpecName: "kube-api-access-7rhfx") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "kube-api-access-7rhfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.946565 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.949652 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:36:51 crc kubenswrapper[4677]: I1203 14:36:51.950938 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7d24039e-bd02-41b9-86b3-1b23b50f6824" (UID: "7d24039e-bd02-41b9-86b3-1b23b50f6824"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.018737 4677 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.018769 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rhfx\" (UniqueName: \"kubernetes.io/projected/7d24039e-bd02-41b9-86b3-1b23b50f6824-kube-api-access-7rhfx\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.018779 4677 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.018788 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d24039e-bd02-41b9-86b3-1b23b50f6824-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.359481 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" event={"ID":"7d24039e-bd02-41b9-86b3-1b23b50f6824","Type":"ContainerDied","Data":"294a7cc4102e6db1d0c097128920bd5692a3111f9a5f096ad761d59ba5621417"} Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.359536 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="294a7cc4102e6db1d0c097128920bd5692a3111f9a5f096ad761d59ba5621417" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.359614 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hcrwl" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.455400 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97"] Dec 03 14:36:52 crc kubenswrapper[4677]: E1203 14:36:52.456100 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d24039e-bd02-41b9-86b3-1b23b50f6824" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.456166 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d24039e-bd02-41b9-86b3-1b23b50f6824" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:36:52 crc kubenswrapper[4677]: E1203 14:36:52.456196 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="extract-utilities" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.456206 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="extract-utilities" Dec 03 14:36:52 crc kubenswrapper[4677]: E1203 14:36:52.456223 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="registry-server" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.456231 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="registry-server" Dec 03 14:36:52 crc kubenswrapper[4677]: E1203 14:36:52.456245 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="extract-content" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.456252 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="extract-content" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.456527 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="04155b7a-f655-4013-a6d2-a06c9a3b3d80" containerName="registry-server" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.456565 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d24039e-bd02-41b9-86b3-1b23b50f6824" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.457460 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.462940 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.463249 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.463460 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qlmtv" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.463494 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.463687 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.471298 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97"] Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.632257 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.632571 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.632631 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.632673 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.632708 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.632776 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.632802 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b59s6\" (UniqueName: \"kubernetes.io/projected/34703414-a555-4edf-addc-8c36c829718b-kube-api-access-b59s6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.734446 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b59s6\" (UniqueName: \"kubernetes.io/projected/34703414-a555-4edf-addc-8c36c829718b-kube-api-access-b59s6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.734515 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.734597 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.734680 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.734735 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.734782 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.734880 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.741366 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.741397 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.741583 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.742122 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.744649 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.755852 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.757071 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b59s6\" (UniqueName: \"kubernetes.io/projected/34703414-a555-4edf-addc-8c36c829718b-kube-api-access-b59s6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6c97\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:52 crc kubenswrapper[4677]: I1203 14:36:52.775303 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:36:53 crc kubenswrapper[4677]: I1203 14:36:53.370674 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:36:53 crc kubenswrapper[4677]: I1203 14:36:53.370756 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97"] Dec 03 14:36:54 crc kubenswrapper[4677]: I1203 14:36:54.385248 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" event={"ID":"34703414-a555-4edf-addc-8c36c829718b","Type":"ContainerStarted","Data":"35e4d8337876b390e0abd15ce68f480f6832f668fd3b8495a5dd104ebc27b799"} Dec 03 14:36:54 crc kubenswrapper[4677]: I1203 14:36:54.385606 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" event={"ID":"34703414-a555-4edf-addc-8c36c829718b","Type":"ContainerStarted","Data":"c23393af470e4d0437cae92402d460fb5bf3a6ec8d3eba7c1627942a243dc3ee"} Dec 03 14:37:08 crc kubenswrapper[4677]: I1203 14:37:08.437022 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:37:08 crc kubenswrapper[4677]: I1203 14:37:08.437656 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:37:08 crc kubenswrapper[4677]: I1203 14:37:08.437714 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:37:08 crc kubenswrapper[4677]: I1203 14:37:08.438657 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:37:08 crc kubenswrapper[4677]: I1203 14:37:08.438727 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" gracePeriod=600 Dec 03 14:37:08 crc kubenswrapper[4677]: E1203 14:37:08.562255 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:37:09 crc kubenswrapper[4677]: I1203 14:37:09.552034 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" exitCode=0 Dec 03 14:37:09 crc kubenswrapper[4677]: I1203 14:37:09.552135 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4"} Dec 03 14:37:09 crc kubenswrapper[4677]: I1203 14:37:09.552568 4677 scope.go:117] "RemoveContainer" containerID="2d41f9dabdf4731fdfeb081c4a938a6ca90a305c55b57b82cac90a2c2d9cb5c5" Dec 03 14:37:09 crc kubenswrapper[4677]: I1203 14:37:09.554160 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:37:09 crc kubenswrapper[4677]: E1203 14:37:09.554933 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:37:09 crc kubenswrapper[4677]: I1203 14:37:09.586253 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" podStartSLOduration=17.02541344 podStartE2EDuration="17.586214881s" podCreationTimestamp="2025-12-03 14:36:52 +0000 UTC" firstStartedPulling="2025-12-03 14:36:53.370341459 +0000 UTC m=+3004.116673914" lastFinishedPulling="2025-12-03 14:36:53.9311429 +0000 UTC m=+3004.677475355" observedRunningTime="2025-12-03 14:36:54.410488698 +0000 UTC m=+3005.156821163" watchObservedRunningTime="2025-12-03 14:37:09.586214881 +0000 UTC m=+3020.332547336" Dec 03 14:37:24 crc kubenswrapper[4677]: I1203 14:37:24.976792 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:37:24 crc kubenswrapper[4677]: E1203 14:37:24.977621 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:37:38 crc kubenswrapper[4677]: I1203 14:37:38.976250 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:37:38 crc kubenswrapper[4677]: E1203 14:37:38.977186 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:37:49 crc kubenswrapper[4677]: I1203 14:37:49.988628 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:37:49 crc kubenswrapper[4677]: E1203 14:37:49.989646 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:38:01 crc kubenswrapper[4677]: I1203 14:38:01.976143 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:38:01 crc kubenswrapper[4677]: E1203 14:38:01.976874 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:38:12 crc kubenswrapper[4677]: I1203 14:38:12.976194 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:38:12 crc kubenswrapper[4677]: E1203 14:38:12.977053 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:38:27 crc kubenswrapper[4677]: I1203 14:38:27.976541 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:38:27 crc kubenswrapper[4677]: E1203 14:38:27.977422 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:38:38 crc kubenswrapper[4677]: I1203 14:38:38.976131 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:38:38 crc kubenswrapper[4677]: E1203 14:38:38.976938 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:38:49 crc kubenswrapper[4677]: I1203 14:38:49.982162 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:38:49 crc kubenswrapper[4677]: E1203 14:38:49.982848 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:39:01 crc kubenswrapper[4677]: I1203 14:39:01.977059 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:39:01 crc kubenswrapper[4677]: E1203 14:39:01.977738 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:39:13 crc kubenswrapper[4677]: I1203 14:39:13.834235 4677 generic.go:334] "Generic (PLEG): container finished" podID="34703414-a555-4edf-addc-8c36c829718b" containerID="35e4d8337876b390e0abd15ce68f480f6832f668fd3b8495a5dd104ebc27b799" exitCode=0 Dec 03 14:39:13 crc kubenswrapper[4677]: I1203 14:39:13.834294 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" event={"ID":"34703414-a555-4edf-addc-8c36c829718b","Type":"ContainerDied","Data":"35e4d8337876b390e0abd15ce68f480f6832f668fd3b8495a5dd104ebc27b799"} Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.404102 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.518232 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ssh-key\") pod \"34703414-a555-4edf-addc-8c36c829718b\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.518716 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-0\") pod \"34703414-a555-4edf-addc-8c36c829718b\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.519376 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-telemetry-combined-ca-bundle\") pod \"34703414-a555-4edf-addc-8c36c829718b\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.519449 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-inventory\") pod \"34703414-a555-4edf-addc-8c36c829718b\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.519479 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-2\") pod \"34703414-a555-4edf-addc-8c36c829718b\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.519500 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b59s6\" (UniqueName: \"kubernetes.io/projected/34703414-a555-4edf-addc-8c36c829718b-kube-api-access-b59s6\") pod \"34703414-a555-4edf-addc-8c36c829718b\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.519520 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-1\") pod \"34703414-a555-4edf-addc-8c36c829718b\" (UID: \"34703414-a555-4edf-addc-8c36c829718b\") " Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.527403 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "34703414-a555-4edf-addc-8c36c829718b" (UID: "34703414-a555-4edf-addc-8c36c829718b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.527555 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34703414-a555-4edf-addc-8c36c829718b-kube-api-access-b59s6" (OuterVolumeSpecName: "kube-api-access-b59s6") pod "34703414-a555-4edf-addc-8c36c829718b" (UID: "34703414-a555-4edf-addc-8c36c829718b"). InnerVolumeSpecName "kube-api-access-b59s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.559519 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-inventory" (OuterVolumeSpecName: "inventory") pod "34703414-a555-4edf-addc-8c36c829718b" (UID: "34703414-a555-4edf-addc-8c36c829718b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.560817 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "34703414-a555-4edf-addc-8c36c829718b" (UID: "34703414-a555-4edf-addc-8c36c829718b"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.564087 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "34703414-a555-4edf-addc-8c36c829718b" (UID: "34703414-a555-4edf-addc-8c36c829718b"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.569484 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "34703414-a555-4edf-addc-8c36c829718b" (UID: "34703414-a555-4edf-addc-8c36c829718b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.577165 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "34703414-a555-4edf-addc-8c36c829718b" (UID: "34703414-a555-4edf-addc-8c36c829718b"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.624101 4677 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.624144 4677 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.624160 4677 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.624174 4677 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.624186 4677 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.624201 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b59s6\" (UniqueName: \"kubernetes.io/projected/34703414-a555-4edf-addc-8c36c829718b-kube-api-access-b59s6\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.624216 4677 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/34703414-a555-4edf-addc-8c36c829718b-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.858544 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" event={"ID":"34703414-a555-4edf-addc-8c36c829718b","Type":"ContainerDied","Data":"c23393af470e4d0437cae92402d460fb5bf3a6ec8d3eba7c1627942a243dc3ee"} Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.858583 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6c97" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.858606 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c23393af470e4d0437cae92402d460fb5bf3a6ec8d3eba7c1627942a243dc3ee" Dec 03 14:39:15 crc kubenswrapper[4677]: I1203 14:39:15.976858 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:39:15 crc kubenswrapper[4677]: E1203 14:39:15.977255 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:39:30 crc kubenswrapper[4677]: I1203 14:39:30.976712 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:39:30 crc kubenswrapper[4677]: E1203 14:39:30.978042 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:39:45 crc kubenswrapper[4677]: I1203 14:39:45.976127 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:39:45 crc kubenswrapper[4677]: E1203 14:39:45.976975 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:39:53 crc kubenswrapper[4677]: I1203 14:39:53.901838 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 03 14:39:53 crc kubenswrapper[4677]: E1203 14:39:53.903611 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34703414-a555-4edf-addc-8c36c829718b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 14:39:53 crc kubenswrapper[4677]: I1203 14:39:53.903637 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="34703414-a555-4edf-addc-8c36c829718b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 14:39:53 crc kubenswrapper[4677]: I1203 14:39:53.904036 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="34703414-a555-4edf-addc-8c36c829718b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 14:39:53 crc kubenswrapper[4677]: I1203 14:39:53.905602 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 14:39:53 crc kubenswrapper[4677]: I1203 14:39:53.909575 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 03 14:39:53 crc kubenswrapper[4677]: I1203 14:39:53.925337 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.094205 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.096721 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.104512 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-2-config-data" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.107399 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj9v2\" (UniqueName: \"kubernetes.io/projected/b663c30a-e2f1-474f-9a25-22d9ab5c571c-kube-api-access-wj9v2\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.107525 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-run\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.107608 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.107731 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109039 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109133 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-sys\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109213 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109283 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-dev\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109405 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-config-data\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109481 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109601 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.109907 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.110147 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-lib-modules\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.110251 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-scripts\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.110413 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.139655 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.164286 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-0"] Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.166274 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.169367 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-config-data" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.196927 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212601 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212650 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212685 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212709 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212728 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212755 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212774 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-sys\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212803 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-dev\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212837 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212895 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-config-data\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212914 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.212934 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213003 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213057 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213086 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213106 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213132 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-lib-modules\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213152 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-scripts\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213176 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213209 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213229 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213264 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj9v2\" (UniqueName: \"kubernetes.io/projected/b663c30a-e2f1-474f-9a25-22d9ab5c571c-kube-api-access-wj9v2\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213296 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213319 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-run\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213346 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213395 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213427 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfv54\" (UniqueName: \"kubernetes.io/projected/387b7e47-02b4-463b-a669-3cc5edc6fee6-kube-api-access-sfv54\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213452 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213478 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.213509 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.214304 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-sys\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.214683 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.214769 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-run\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.214870 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.214911 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.215131 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-lib-modules\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.215922 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.217168 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-dev\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.217305 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.222351 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-scripts\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.225942 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b663c30a-e2f1-474f-9a25-22d9ab5c571c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.226747 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.227263 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-config-data\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.235957 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b663c30a-e2f1-474f-9a25-22d9ab5c571c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.239067 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj9v2\" (UniqueName: \"kubernetes.io/projected/b663c30a-e2f1-474f-9a25-22d9ab5c571c-kube-api-access-wj9v2\") pod \"cinder-backup-0\" (UID: \"b663c30a-e2f1-474f-9a25-22d9ab5c571c\") " pod="openstack/cinder-backup-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315249 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315324 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315365 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315399 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-sys\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315426 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfv54\" (UniqueName: \"kubernetes.io/projected/387b7e47-02b4-463b-a669-3cc5edc6fee6-kube-api-access-sfv54\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315453 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315477 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315506 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315525 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315551 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdfg5\" (UniqueName: \"kubernetes.io/projected/3014c1b9-a994-43e8-b490-f62b76bcf6f6-kube-api-access-qdfg5\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315595 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315622 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315643 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315672 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315694 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315719 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315789 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315834 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-dev\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315861 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315887 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315940 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.315988 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316022 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316054 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316101 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316141 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316163 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316192 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-run\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316217 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316240 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316415 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316775 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.316815 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.317088 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.317255 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.317413 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.317798 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.320021 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.320059 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.320060 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/387b7e47-02b4-463b-a669-3cc5edc6fee6-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.322056 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.322889 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.324292 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.324936 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b7e47-02b4-463b-a669-3cc5edc6fee6-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.341658 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfv54\" (UniqueName: \"kubernetes.io/projected/387b7e47-02b4-463b-a669-3cc5edc6fee6-kube-api-access-sfv54\") pod \"cinder-volume-nfs-2-0\" (UID: \"387b7e47-02b4-463b-a669-3cc5edc6fee6\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418529 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-run\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418592 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418634 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418663 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418701 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-sys\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418733 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418755 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418810 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdfg5\" (UniqueName: \"kubernetes.io/projected/3014c1b9-a994-43e8-b490-f62b76bcf6f6-kube-api-access-qdfg5\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418848 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.418928 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419000 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-dev\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419030 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419107 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419163 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419185 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419340 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419391 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-run\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419432 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.419473 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.420016 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.420066 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-sys\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.423320 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.423528 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.424155 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.424218 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3014c1b9-a994-43e8-b490-f62b76bcf6f6-dev\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.426892 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.426896 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.427017 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.428871 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3014c1b9-a994-43e8-b490-f62b76bcf6f6-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.444683 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.449764 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdfg5\" (UniqueName: \"kubernetes.io/projected/3014c1b9-a994-43e8-b490-f62b76bcf6f6-kube-api-access-qdfg5\") pod \"cinder-volume-nfs-0\" (UID: \"3014c1b9-a994-43e8-b490-f62b76bcf6f6\") " pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.495381 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:54 crc kubenswrapper[4677]: I1203 14:39:54.527393 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 14:39:55 crc kubenswrapper[4677]: I1203 14:39:55.187721 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Dec 03 14:39:55 crc kubenswrapper[4677]: I1203 14:39:55.239001 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"387b7e47-02b4-463b-a669-3cc5edc6fee6","Type":"ContainerStarted","Data":"d3f8624a8c1c9d1983aae668d2c784d73d24b4410353881b23cf454371bed1b5"} Dec 03 14:39:55 crc kubenswrapper[4677]: I1203 14:39:55.277525 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Dec 03 14:39:55 crc kubenswrapper[4677]: W1203 14:39:55.374227 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3014c1b9_a994_43e8_b490_f62b76bcf6f6.slice/crio-4ecd463528cf4a9fb2a717d22f96cef616184d4c7c10d1d182156dad190d9d42 WatchSource:0}: Error finding container 4ecd463528cf4a9fb2a717d22f96cef616184d4c7c10d1d182156dad190d9d42: Status 404 returned error can't find the container with id 4ecd463528cf4a9fb2a717d22f96cef616184d4c7c10d1d182156dad190d9d42 Dec 03 14:39:55 crc kubenswrapper[4677]: W1203 14:39:55.399647 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb663c30a_e2f1_474f_9a25_22d9ab5c571c.slice/crio-3a791abbfe083c8235849c61ceaa7f11369f1ed55733cb294dd13264f092c8c7 WatchSource:0}: Error finding container 3a791abbfe083c8235849c61ceaa7f11369f1ed55733cb294dd13264f092c8c7: Status 404 returned error can't find the container with id 3a791abbfe083c8235849c61ceaa7f11369f1ed55733cb294dd13264f092c8c7 Dec 03 14:39:55 crc kubenswrapper[4677]: I1203 14:39:55.400010 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 14:39:56 crc kubenswrapper[4677]: I1203 14:39:56.294511 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"387b7e47-02b4-463b-a669-3cc5edc6fee6","Type":"ContainerStarted","Data":"f567b6aa00f7693293cbad69fd0bee7b59347bc868363c61cbf99b849f225a1e"} Dec 03 14:39:56 crc kubenswrapper[4677]: I1203 14:39:56.298023 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"b663c30a-e2f1-474f-9a25-22d9ab5c571c","Type":"ContainerStarted","Data":"83640fea53dd272ecc6e94492e1b542cfd3ec8708f52c5a57a9fc7dce8c39bbb"} Dec 03 14:39:56 crc kubenswrapper[4677]: I1203 14:39:56.298073 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"b663c30a-e2f1-474f-9a25-22d9ab5c571c","Type":"ContainerStarted","Data":"3a791abbfe083c8235849c61ceaa7f11369f1ed55733cb294dd13264f092c8c7"} Dec 03 14:39:56 crc kubenswrapper[4677]: I1203 14:39:56.308716 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"3014c1b9-a994-43e8-b490-f62b76bcf6f6","Type":"ContainerStarted","Data":"740c992312b51fb1de916943c61618cc4b9af473e0521a539957289fd8b123ab"} Dec 03 14:39:56 crc kubenswrapper[4677]: I1203 14:39:56.308763 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"3014c1b9-a994-43e8-b490-f62b76bcf6f6","Type":"ContainerStarted","Data":"4ecd463528cf4a9fb2a717d22f96cef616184d4c7c10d1d182156dad190d9d42"} Dec 03 14:39:57 crc kubenswrapper[4677]: I1203 14:39:57.321541 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"387b7e47-02b4-463b-a669-3cc5edc6fee6","Type":"ContainerStarted","Data":"247b3ddca2936b1ad8ecfc48434b7151d224c33fd0e45a2f40f3640ed38bba1c"} Dec 03 14:39:57 crc kubenswrapper[4677]: I1203 14:39:57.324900 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"b663c30a-e2f1-474f-9a25-22d9ab5c571c","Type":"ContainerStarted","Data":"3c5ce0880ad34a8a98a93b0641a2f1ec5ea988813a0bb82604fe172c12ea0691"} Dec 03 14:39:57 crc kubenswrapper[4677]: I1203 14:39:57.328065 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"3014c1b9-a994-43e8-b490-f62b76bcf6f6","Type":"ContainerStarted","Data":"fa871d21744638257b41bc48ab4f9fba75759b7433cc582ae4570eefaaacb1ca"} Dec 03 14:39:57 crc kubenswrapper[4677]: I1203 14:39:57.366862 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-2-0" podStartSLOduration=3.075423867 podStartE2EDuration="3.366834385s" podCreationTimestamp="2025-12-03 14:39:54 +0000 UTC" firstStartedPulling="2025-12-03 14:39:55.19172068 +0000 UTC m=+3185.938053135" lastFinishedPulling="2025-12-03 14:39:55.483131198 +0000 UTC m=+3186.229463653" observedRunningTime="2025-12-03 14:39:57.35102826 +0000 UTC m=+3188.097360735" watchObservedRunningTime="2025-12-03 14:39:57.366834385 +0000 UTC m=+3188.113166830" Dec 03 14:39:57 crc kubenswrapper[4677]: I1203 14:39:57.388908 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=4.046653686 podStartE2EDuration="4.388876782s" podCreationTimestamp="2025-12-03 14:39:53 +0000 UTC" firstStartedPulling="2025-12-03 14:39:55.43090121 +0000 UTC m=+3186.177233665" lastFinishedPulling="2025-12-03 14:39:55.773124306 +0000 UTC m=+3186.519456761" observedRunningTime="2025-12-03 14:39:57.376195172 +0000 UTC m=+3188.122527657" watchObservedRunningTime="2025-12-03 14:39:57.388876782 +0000 UTC m=+3188.135209237" Dec 03 14:39:57 crc kubenswrapper[4677]: I1203 14:39:57.405600 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-0" podStartSLOduration=3.299154793 podStartE2EDuration="3.405578091s" podCreationTimestamp="2025-12-03 14:39:54 +0000 UTC" firstStartedPulling="2025-12-03 14:39:55.37672163 +0000 UTC m=+3186.123054085" lastFinishedPulling="2025-12-03 14:39:55.483144928 +0000 UTC m=+3186.229477383" observedRunningTime="2025-12-03 14:39:57.399473923 +0000 UTC m=+3188.145806398" watchObservedRunningTime="2025-12-03 14:39:57.405578091 +0000 UTC m=+3188.151910546" Dec 03 14:39:57 crc kubenswrapper[4677]: I1203 14:39:57.977114 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:39:57 crc kubenswrapper[4677]: E1203 14:39:57.977695 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:39:59 crc kubenswrapper[4677]: I1203 14:39:59.445464 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:39:59 crc kubenswrapper[4677]: I1203 14:39:59.496272 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-0" Dec 03 14:39:59 crc kubenswrapper[4677]: I1203 14:39:59.528723 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 03 14:40:05 crc kubenswrapper[4677]: I1203 14:40:05.143119 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 03 14:40:05 crc kubenswrapper[4677]: I1203 14:40:05.153427 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-2-0" Dec 03 14:40:05 crc kubenswrapper[4677]: I1203 14:40:05.217857 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-0" Dec 03 14:40:09 crc kubenswrapper[4677]: I1203 14:40:09.985260 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:40:09 crc kubenswrapper[4677]: E1203 14:40:09.986053 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:40:23 crc kubenswrapper[4677]: I1203 14:40:23.976785 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:40:23 crc kubenswrapper[4677]: E1203 14:40:23.977743 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:40:38 crc kubenswrapper[4677]: I1203 14:40:38.976471 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:40:38 crc kubenswrapper[4677]: E1203 14:40:38.977366 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:40:50 crc kubenswrapper[4677]: I1203 14:40:50.000394 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:40:50 crc kubenswrapper[4677]: E1203 14:40:50.001474 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:40:59 crc kubenswrapper[4677]: I1203 14:40:59.751909 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:40:59 crc kubenswrapper[4677]: I1203 14:40:59.752769 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="prometheus" containerID="cri-o://f209203ed74dc0d31ade3ef18bd09b8f182817995dd2b9969d9eea2a59c4efd7" gracePeriod=600 Dec 03 14:40:59 crc kubenswrapper[4677]: I1203 14:40:59.753096 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="config-reloader" containerID="cri-o://f64a2560b1753ecb052a6d47e806c54eeff5b227bd7f96324fd2c656a66c9f84" gracePeriod=600 Dec 03 14:40:59 crc kubenswrapper[4677]: I1203 14:40:59.753086 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="thanos-sidecar" containerID="cri-o://2f216e09675889656977525ecea68681584e97f744c6a2d705d35a6612fa5d86" gracePeriod=600 Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.428534 4677 generic.go:334] "Generic (PLEG): container finished" podID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerID="2f216e09675889656977525ecea68681584e97f744c6a2d705d35a6612fa5d86" exitCode=0 Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.428872 4677 generic.go:334] "Generic (PLEG): container finished" podID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerID="f64a2560b1753ecb052a6d47e806c54eeff5b227bd7f96324fd2c656a66c9f84" exitCode=0 Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.428885 4677 generic.go:334] "Generic (PLEG): container finished" podID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerID="f209203ed74dc0d31ade3ef18bd09b8f182817995dd2b9969d9eea2a59c4efd7" exitCode=0 Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.428757 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerDied","Data":"2f216e09675889656977525ecea68681584e97f744c6a2d705d35a6612fa5d86"} Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.428935 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerDied","Data":"f64a2560b1753ecb052a6d47e806c54eeff5b227bd7f96324fd2c656a66c9f84"} Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.428976 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerDied","Data":"f209203ed74dc0d31ade3ef18bd09b8f182817995dd2b9969d9eea2a59c4efd7"} Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.753882 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.873705 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4mtf\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-kube-api-access-k4mtf\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.873751 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-prometheus-metric-storage-rulefiles-0\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.873938 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874060 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874103 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-thanos-prometheus-http-client-file\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874140 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-secret-combined-ca-bundle\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874185 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-tls-assets\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874238 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config-out\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874384 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874411 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874431 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config\") pod \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\" (UID: \"edfdba14-4c6d-4708-b024-2f5c7bc2f01b\") " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.874738 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.875193 4677 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.882855 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.883548 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.884268 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.884783 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.885312 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config" (OuterVolumeSpecName: "config") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.885403 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config-out" (OuterVolumeSpecName: "config-out") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.886162 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.886513 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-kube-api-access-k4mtf" (OuterVolumeSpecName: "kube-api-access-k4mtf") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "kube-api-access-k4mtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.924059 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977290 4677 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977332 4677 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977348 4677 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977364 4677 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977380 4677 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977393 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4mtf\" (UniqueName: \"kubernetes.io/projected/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-kube-api-access-k4mtf\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977427 4677 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") on node \"crc\" " Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977441 4677 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.977457 4677 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:00 crc kubenswrapper[4677]: I1203 14:41:00.998114 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config" (OuterVolumeSpecName: "web-config") pod "edfdba14-4c6d-4708-b024-2f5c7bc2f01b" (UID: "edfdba14-4c6d-4708-b024-2f5c7bc2f01b"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.002363 4677 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.002503 4677 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208") on node "crc" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.080754 4677 reconciler_common.go:293] "Volume detached for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.080783 4677 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/edfdba14-4c6d-4708-b024-2f5c7bc2f01b-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.444396 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"edfdba14-4c6d-4708-b024-2f5c7bc2f01b","Type":"ContainerDied","Data":"f8276e27efd4f4de3a99f9723cf8afde1aaf1cf37781d7cb7413fd4b878f591c"} Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.444461 4677 scope.go:117] "RemoveContainer" containerID="2f216e09675889656977525ecea68681584e97f744c6a2d705d35a6612fa5d86" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.444559 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.477362 4677 scope.go:117] "RemoveContainer" containerID="f64a2560b1753ecb052a6d47e806c54eeff5b227bd7f96324fd2c656a66c9f84" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.496414 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.510504 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.518391 4677 scope.go:117] "RemoveContainer" containerID="f209203ed74dc0d31ade3ef18bd09b8f182817995dd2b9969d9eea2a59c4efd7" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525033 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:41:01 crc kubenswrapper[4677]: E1203 14:41:01.525526 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="prometheus" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525553 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="prometheus" Dec 03 14:41:01 crc kubenswrapper[4677]: E1203 14:41:01.525564 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="thanos-sidecar" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525573 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="thanos-sidecar" Dec 03 14:41:01 crc kubenswrapper[4677]: E1203 14:41:01.525587 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="init-config-reloader" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525596 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="init-config-reloader" Dec 03 14:41:01 crc kubenswrapper[4677]: E1203 14:41:01.525622 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="config-reloader" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525630 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="config-reloader" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525876 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="thanos-sidecar" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525897 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="prometheus" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.525909 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" containerName="config-reloader" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.527807 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.531304 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.531326 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.541665 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mq9g8" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.543417 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.550266 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.551335 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.555885 4677 scope.go:117] "RemoveContainer" containerID="131b5669e7bdb8fec5a14c35b257c557a9085a95a7fe9a66bf08a18a6b349756" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.556641 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595354 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595420 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595469 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-config\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595496 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595531 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/429e5aed-7233-4056-a808-9afc77fa5011-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595561 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/429e5aed-7233-4056-a808-9afc77fa5011-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595594 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595626 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595650 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595671 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/429e5aed-7233-4056-a808-9afc77fa5011-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.595797 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l99v\" (UniqueName: \"kubernetes.io/projected/429e5aed-7233-4056-a808-9afc77fa5011-kube-api-access-2l99v\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698574 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l99v\" (UniqueName: \"kubernetes.io/projected/429e5aed-7233-4056-a808-9afc77fa5011-kube-api-access-2l99v\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698652 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698693 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698768 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-config\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698799 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698848 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/429e5aed-7233-4056-a808-9afc77fa5011-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698891 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/429e5aed-7233-4056-a808-9afc77fa5011-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.698942 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.699013 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.699041 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.699066 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/429e5aed-7233-4056-a808-9afc77fa5011-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.702170 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/429e5aed-7233-4056-a808-9afc77fa5011-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.703306 4677 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.703347 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b3062aca7e122ea7b9aa119622cdce8d97263fccd877be8c6301bd417d0b47b0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.705307 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/429e5aed-7233-4056-a808-9afc77fa5011-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.705629 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.705668 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/429e5aed-7233-4056-a808-9afc77fa5011-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.707485 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.707516 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.708585 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-config\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.709388 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.716346 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/429e5aed-7233-4056-a808-9afc77fa5011-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.718772 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l99v\" (UniqueName: \"kubernetes.io/projected/429e5aed-7233-4056-a808-9afc77fa5011-kube-api-access-2l99v\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.750000 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505eb4a1-f3de-4b2a-9b2a-94d650b03208\") pod \"prometheus-metric-storage-0\" (UID: \"429e5aed-7233-4056-a808-9afc77fa5011\") " pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.852412 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:01 crc kubenswrapper[4677]: I1203 14:41:01.995835 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edfdba14-4c6d-4708-b024-2f5c7bc2f01b" path="/var/lib/kubelet/pods/edfdba14-4c6d-4708-b024-2f5c7bc2f01b/volumes" Dec 03 14:41:02 crc kubenswrapper[4677]: I1203 14:41:02.361176 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 14:41:02 crc kubenswrapper[4677]: I1203 14:41:02.456786 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"429e5aed-7233-4056-a808-9afc77fa5011","Type":"ContainerStarted","Data":"0b6b0a554d0df9e79123a000d334b4cc746355a2411f5c3717a095c8d1e03cdd"} Dec 03 14:41:04 crc kubenswrapper[4677]: I1203 14:41:04.976312 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:41:04 crc kubenswrapper[4677]: E1203 14:41:04.977144 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:41:06 crc kubenswrapper[4677]: I1203 14:41:06.500503 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"429e5aed-7233-4056-a808-9afc77fa5011","Type":"ContainerStarted","Data":"44f36184945721ac0316a35a6a898e200b2560dad358b5b11ae77accee021701"} Dec 03 14:41:14 crc kubenswrapper[4677]: I1203 14:41:14.598893 4677 generic.go:334] "Generic (PLEG): container finished" podID="429e5aed-7233-4056-a808-9afc77fa5011" containerID="44f36184945721ac0316a35a6a898e200b2560dad358b5b11ae77accee021701" exitCode=0 Dec 03 14:41:14 crc kubenswrapper[4677]: I1203 14:41:14.598997 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"429e5aed-7233-4056-a808-9afc77fa5011","Type":"ContainerDied","Data":"44f36184945721ac0316a35a6a898e200b2560dad358b5b11ae77accee021701"} Dec 03 14:41:15 crc kubenswrapper[4677]: I1203 14:41:15.613315 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"429e5aed-7233-4056-a808-9afc77fa5011","Type":"ContainerStarted","Data":"06ac576ecb7605988ee83ce617eb992c92525fcb660456558e4985e6b6c94d3d"} Dec 03 14:41:19 crc kubenswrapper[4677]: I1203 14:41:19.985068 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:41:19 crc kubenswrapper[4677]: E1203 14:41:19.986213 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:41:20 crc kubenswrapper[4677]: I1203 14:41:20.662936 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"429e5aed-7233-4056-a808-9afc77fa5011","Type":"ContainerStarted","Data":"8a4b40871e814c1cf83100c68a3122ddd24f98ee9df6e14df1d924d1ac8454ef"} Dec 03 14:41:20 crc kubenswrapper[4677]: I1203 14:41:20.662993 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"429e5aed-7233-4056-a808-9afc77fa5011","Type":"ContainerStarted","Data":"a6fe5b8d5bdb37af3dbf88820535220e9aa2b9c39925a84dad7cd1c45a29fd57"} Dec 03 14:41:20 crc kubenswrapper[4677]: I1203 14:41:20.713418 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.71339832 podStartE2EDuration="19.71339832s" podCreationTimestamp="2025-12-03 14:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:41:20.693628919 +0000 UTC m=+3271.439961384" watchObservedRunningTime="2025-12-03 14:41:20.71339832 +0000 UTC m=+3271.459730785" Dec 03 14:41:21 crc kubenswrapper[4677]: I1203 14:41:21.853200 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:31 crc kubenswrapper[4677]: I1203 14:41:31.853606 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:31 crc kubenswrapper[4677]: I1203 14:41:31.863525 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:32 crc kubenswrapper[4677]: I1203 14:41:32.794256 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 14:41:33 crc kubenswrapper[4677]: I1203 14:41:33.977004 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:41:33 crc kubenswrapper[4677]: E1203 14:41:33.977590 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.180444 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.182498 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.186385 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.186501 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.186899 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-dbb9m" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.187391 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.200648 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278263 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gphvm\" (UniqueName: \"kubernetes.io/projected/e4245f29-1973-483f-ad9c-ab450ae9f6d6-kube-api-access-gphvm\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278352 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278389 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e4245f29-1973-483f-ad9c-ab450ae9f6d6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278479 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278527 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278573 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278622 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e4245f29-1973-483f-ad9c-ab450ae9f6d6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278664 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e4245f29-1973-483f-ad9c-ab450ae9f6d6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.278693 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4245f29-1973-483f-ad9c-ab450ae9f6d6-config-data\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.380990 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gphvm\" (UniqueName: \"kubernetes.io/projected/e4245f29-1973-483f-ad9c-ab450ae9f6d6-kube-api-access-gphvm\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381337 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381363 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e4245f29-1973-483f-ad9c-ab450ae9f6d6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381437 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381557 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381670 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381716 4677 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381739 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e4245f29-1973-483f-ad9c-ab450ae9f6d6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381799 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e4245f29-1973-483f-ad9c-ab450ae9f6d6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.381842 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4245f29-1973-483f-ad9c-ab450ae9f6d6-config-data\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.382234 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e4245f29-1973-483f-ad9c-ab450ae9f6d6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.383404 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4245f29-1973-483f-ad9c-ab450ae9f6d6-config-data\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.384184 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e4245f29-1973-483f-ad9c-ab450ae9f6d6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.384428 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e4245f29-1973-483f-ad9c-ab450ae9f6d6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.390002 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.391257 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.395039 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4245f29-1973-483f-ad9c-ab450ae9f6d6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.402517 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gphvm\" (UniqueName: \"kubernetes.io/projected/e4245f29-1973-483f-ad9c-ab450ae9f6d6-kube-api-access-gphvm\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.422399 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"e4245f29-1973-483f-ad9c-ab450ae9f6d6\") " pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.503412 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 14:41:43 crc kubenswrapper[4677]: I1203 14:41:43.993267 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 14:41:44 crc kubenswrapper[4677]: I1203 14:41:44.906057 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"e4245f29-1973-483f-ad9c-ab450ae9f6d6","Type":"ContainerStarted","Data":"fb0867c6cd07a1dc59a132a38925a3db7a6804084a7d0474592747713f02d081"} Dec 03 14:41:46 crc kubenswrapper[4677]: I1203 14:41:46.977242 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:41:46 crc kubenswrapper[4677]: E1203 14:41:46.977809 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:41:54 crc kubenswrapper[4677]: I1203 14:41:54.504329 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 14:41:56 crc kubenswrapper[4677]: I1203 14:41:56.033498 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"e4245f29-1973-483f-ad9c-ab450ae9f6d6","Type":"ContainerStarted","Data":"7260e9eacbabb365aa78c32a02765b225fcb2d37166918e0d60aae28014a38b1"} Dec 03 14:41:56 crc kubenswrapper[4677]: I1203 14:41:56.054157 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.539513411 podStartE2EDuration="14.054136024s" podCreationTimestamp="2025-12-03 14:41:42 +0000 UTC" firstStartedPulling="2025-12-03 14:41:43.986434477 +0000 UTC m=+3294.732766932" lastFinishedPulling="2025-12-03 14:41:54.50105708 +0000 UTC m=+3305.247389545" observedRunningTime="2025-12-03 14:41:56.051937014 +0000 UTC m=+3306.798269489" watchObservedRunningTime="2025-12-03 14:41:56.054136024 +0000 UTC m=+3306.800468479" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.248391 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9c8cb"] Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.251107 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.265085 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9c8cb"] Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.316261 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-utilities\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.316740 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-catalog-content\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.316794 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8qmk\" (UniqueName: \"kubernetes.io/projected/81a2f4c2-4476-41d2-8723-c86767406048-kube-api-access-j8qmk\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.419051 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-utilities\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.419192 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-catalog-content\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.419217 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8qmk\" (UniqueName: \"kubernetes.io/projected/81a2f4c2-4476-41d2-8723-c86767406048-kube-api-access-j8qmk\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.420065 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-utilities\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.420292 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-catalog-content\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.442362 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8qmk\" (UniqueName: \"kubernetes.io/projected/81a2f4c2-4476-41d2-8723-c86767406048-kube-api-access-j8qmk\") pod \"community-operators-9c8cb\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:57 crc kubenswrapper[4677]: I1203 14:41:57.580915 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:41:58 crc kubenswrapper[4677]: I1203 14:41:58.182436 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9c8cb"] Dec 03 14:41:58 crc kubenswrapper[4677]: W1203 14:41:58.191273 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81a2f4c2_4476_41d2_8723_c86767406048.slice/crio-bde968cfa5420246f64cfc1b85f87768ac833c2cf568ba039ed79ae91cdbf626 WatchSource:0}: Error finding container bde968cfa5420246f64cfc1b85f87768ac833c2cf568ba039ed79ae91cdbf626: Status 404 returned error can't find the container with id bde968cfa5420246f64cfc1b85f87768ac833c2cf568ba039ed79ae91cdbf626 Dec 03 14:41:58 crc kubenswrapper[4677]: I1203 14:41:58.977641 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:41:58 crc kubenswrapper[4677]: E1203 14:41:58.978305 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:41:59 crc kubenswrapper[4677]: I1203 14:41:59.073987 4677 generic.go:334] "Generic (PLEG): container finished" podID="81a2f4c2-4476-41d2-8723-c86767406048" containerID="3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06" exitCode=0 Dec 03 14:41:59 crc kubenswrapper[4677]: I1203 14:41:59.074037 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8cb" event={"ID":"81a2f4c2-4476-41d2-8723-c86767406048","Type":"ContainerDied","Data":"3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06"} Dec 03 14:41:59 crc kubenswrapper[4677]: I1203 14:41:59.074065 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8cb" event={"ID":"81a2f4c2-4476-41d2-8723-c86767406048","Type":"ContainerStarted","Data":"bde968cfa5420246f64cfc1b85f87768ac833c2cf568ba039ed79ae91cdbf626"} Dec 03 14:41:59 crc kubenswrapper[4677]: I1203 14:41:59.076612 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:42:01 crc kubenswrapper[4677]: I1203 14:42:01.095747 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8cb" event={"ID":"81a2f4c2-4476-41d2-8723-c86767406048","Type":"ContainerStarted","Data":"bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80"} Dec 03 14:42:02 crc kubenswrapper[4677]: I1203 14:42:02.108625 4677 generic.go:334] "Generic (PLEG): container finished" podID="81a2f4c2-4476-41d2-8723-c86767406048" containerID="bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80" exitCode=0 Dec 03 14:42:02 crc kubenswrapper[4677]: I1203 14:42:02.108703 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8cb" event={"ID":"81a2f4c2-4476-41d2-8723-c86767406048","Type":"ContainerDied","Data":"bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80"} Dec 03 14:42:03 crc kubenswrapper[4677]: I1203 14:42:03.124997 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8cb" event={"ID":"81a2f4c2-4476-41d2-8723-c86767406048","Type":"ContainerStarted","Data":"54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1"} Dec 03 14:42:03 crc kubenswrapper[4677]: I1203 14:42:03.158218 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9c8cb" podStartSLOduration=2.711634334 podStartE2EDuration="6.158194183s" podCreationTimestamp="2025-12-03 14:41:57 +0000 UTC" firstStartedPulling="2025-12-03 14:41:59.076233582 +0000 UTC m=+3309.822566037" lastFinishedPulling="2025-12-03 14:42:02.522793431 +0000 UTC m=+3313.269125886" observedRunningTime="2025-12-03 14:42:03.146489553 +0000 UTC m=+3313.892822098" watchObservedRunningTime="2025-12-03 14:42:03.158194183 +0000 UTC m=+3313.904526638" Dec 03 14:42:07 crc kubenswrapper[4677]: I1203 14:42:07.582117 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:42:07 crc kubenswrapper[4677]: I1203 14:42:07.583163 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:42:07 crc kubenswrapper[4677]: I1203 14:42:07.645367 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:42:08 crc kubenswrapper[4677]: I1203 14:42:08.218979 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:42:08 crc kubenswrapper[4677]: I1203 14:42:08.272842 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9c8cb"] Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.186419 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9c8cb" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="registry-server" containerID="cri-o://54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1" gracePeriod=2 Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.723449 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.752908 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-catalog-content\") pod \"81a2f4c2-4476-41d2-8723-c86767406048\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.753019 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-utilities\") pod \"81a2f4c2-4476-41d2-8723-c86767406048\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.753161 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8qmk\" (UniqueName: \"kubernetes.io/projected/81a2f4c2-4476-41d2-8723-c86767406048-kube-api-access-j8qmk\") pod \"81a2f4c2-4476-41d2-8723-c86767406048\" (UID: \"81a2f4c2-4476-41d2-8723-c86767406048\") " Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.755330 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-utilities" (OuterVolumeSpecName: "utilities") pod "81a2f4c2-4476-41d2-8723-c86767406048" (UID: "81a2f4c2-4476-41d2-8723-c86767406048"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.762275 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81a2f4c2-4476-41d2-8723-c86767406048-kube-api-access-j8qmk" (OuterVolumeSpecName: "kube-api-access-j8qmk") pod "81a2f4c2-4476-41d2-8723-c86767406048" (UID: "81a2f4c2-4476-41d2-8723-c86767406048"). InnerVolumeSpecName "kube-api-access-j8qmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.838702 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81a2f4c2-4476-41d2-8723-c86767406048" (UID: "81a2f4c2-4476-41d2-8723-c86767406048"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.855579 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8qmk\" (UniqueName: \"kubernetes.io/projected/81a2f4c2-4476-41d2-8723-c86767406048-kube-api-access-j8qmk\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.855871 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.855881 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81a2f4c2-4476-41d2-8723-c86767406048-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:10 crc kubenswrapper[4677]: I1203 14:42:10.977044 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.200528 4677 generic.go:334] "Generic (PLEG): container finished" podID="81a2f4c2-4476-41d2-8723-c86767406048" containerID="54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1" exitCode=0 Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.200600 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9c8cb" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.200599 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8cb" event={"ID":"81a2f4c2-4476-41d2-8723-c86767406048","Type":"ContainerDied","Data":"54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1"} Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.200670 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9c8cb" event={"ID":"81a2f4c2-4476-41d2-8723-c86767406048","Type":"ContainerDied","Data":"bde968cfa5420246f64cfc1b85f87768ac833c2cf568ba039ed79ae91cdbf626"} Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.200691 4677 scope.go:117] "RemoveContainer" containerID="54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.237605 4677 scope.go:117] "RemoveContainer" containerID="bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.242705 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9c8cb"] Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.259303 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9c8cb"] Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.321531 4677 scope.go:117] "RemoveContainer" containerID="3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.374868 4677 scope.go:117] "RemoveContainer" containerID="54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1" Dec 03 14:42:11 crc kubenswrapper[4677]: E1203 14:42:11.375364 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1\": container with ID starting with 54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1 not found: ID does not exist" containerID="54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.375403 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1"} err="failed to get container status \"54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1\": rpc error: code = NotFound desc = could not find container \"54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1\": container with ID starting with 54df7ac780a0c9393ca465aff98f0b2cc0d7022df16e8a92f3a2a3ee88a6cae1 not found: ID does not exist" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.375426 4677 scope.go:117] "RemoveContainer" containerID="bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80" Dec 03 14:42:11 crc kubenswrapper[4677]: E1203 14:42:11.375979 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80\": container with ID starting with bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80 not found: ID does not exist" containerID="bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.376025 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80"} err="failed to get container status \"bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80\": rpc error: code = NotFound desc = could not find container \"bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80\": container with ID starting with bd3cceae9add9c1e408db789ae0a6653319378b22473b4c3125c5527c1ec7c80 not found: ID does not exist" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.376055 4677 scope.go:117] "RemoveContainer" containerID="3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06" Dec 03 14:42:11 crc kubenswrapper[4677]: E1203 14:42:11.376370 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06\": container with ID starting with 3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06 not found: ID does not exist" containerID="3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.376399 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06"} err="failed to get container status \"3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06\": rpc error: code = NotFound desc = could not find container \"3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06\": container with ID starting with 3405a62318ec9f55fa8a50c3d0c50c1f90c3620fe9c34231d4e3ebfa16c10a06 not found: ID does not exist" Dec 03 14:42:11 crc kubenswrapper[4677]: I1203 14:42:11.992744 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81a2f4c2-4476-41d2-8723-c86767406048" path="/var/lib/kubelet/pods/81a2f4c2-4476-41d2-8723-c86767406048/volumes" Dec 03 14:42:12 crc kubenswrapper[4677]: I1203 14:42:12.213410 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"46b3c7a220336ed2f24070b4e085e4e629b1f44e1f35f46cb817a51b75e6f587"} Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.507363 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zk6bg"] Dec 03 14:44:12 crc kubenswrapper[4677]: E1203 14:44:12.509727 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="registry-server" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.509842 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="registry-server" Dec 03 14:44:12 crc kubenswrapper[4677]: E1203 14:44:12.509907 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="extract-content" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.509986 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="extract-content" Dec 03 14:44:12 crc kubenswrapper[4677]: E1203 14:44:12.510072 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="extract-utilities" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.510129 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="extract-utilities" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.510547 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="81a2f4c2-4476-41d2-8723-c86767406048" containerName="registry-server" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.515911 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.523873 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zk6bg"] Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.613280 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r8ng\" (UniqueName: \"kubernetes.io/projected/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-kube-api-access-6r8ng\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.613424 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-catalog-content\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.613509 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-utilities\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.715990 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r8ng\" (UniqueName: \"kubernetes.io/projected/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-kube-api-access-6r8ng\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.716133 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-catalog-content\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.716240 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-utilities\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.716724 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-utilities\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.717258 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-catalog-content\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.748112 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r8ng\" (UniqueName: \"kubernetes.io/projected/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-kube-api-access-6r8ng\") pod \"redhat-operators-zk6bg\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:12 crc kubenswrapper[4677]: I1203 14:44:12.850540 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:13 crc kubenswrapper[4677]: I1203 14:44:13.370259 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zk6bg"] Dec 03 14:44:13 crc kubenswrapper[4677]: W1203 14:44:13.383740 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4e8d38c_6293_4531_a8d2_1c8c10848ff4.slice/crio-0111782ec6ec9fbc7d5d4346e869ee410c6da6f95a88b6da0c60ce69447daafe WatchSource:0}: Error finding container 0111782ec6ec9fbc7d5d4346e869ee410c6da6f95a88b6da0c60ce69447daafe: Status 404 returned error can't find the container with id 0111782ec6ec9fbc7d5d4346e869ee410c6da6f95a88b6da0c60ce69447daafe Dec 03 14:44:13 crc kubenswrapper[4677]: I1203 14:44:13.508848 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk6bg" event={"ID":"e4e8d38c-6293-4531-a8d2-1c8c10848ff4","Type":"ContainerStarted","Data":"0111782ec6ec9fbc7d5d4346e869ee410c6da6f95a88b6da0c60ce69447daafe"} Dec 03 14:44:14 crc kubenswrapper[4677]: I1203 14:44:14.517926 4677 generic.go:334] "Generic (PLEG): container finished" podID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerID="e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a" exitCode=0 Dec 03 14:44:14 crc kubenswrapper[4677]: I1203 14:44:14.518133 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk6bg" event={"ID":"e4e8d38c-6293-4531-a8d2-1c8c10848ff4","Type":"ContainerDied","Data":"e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a"} Dec 03 14:44:17 crc kubenswrapper[4677]: I1203 14:44:17.695405 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk6bg" event={"ID":"e4e8d38c-6293-4531-a8d2-1c8c10848ff4","Type":"ContainerStarted","Data":"b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741"} Dec 03 14:44:22 crc kubenswrapper[4677]: I1203 14:44:22.768220 4677 generic.go:334] "Generic (PLEG): container finished" podID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerID="b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741" exitCode=0 Dec 03 14:44:22 crc kubenswrapper[4677]: I1203 14:44:22.768451 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk6bg" event={"ID":"e4e8d38c-6293-4531-a8d2-1c8c10848ff4","Type":"ContainerDied","Data":"b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741"} Dec 03 14:44:23 crc kubenswrapper[4677]: I1203 14:44:23.785088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk6bg" event={"ID":"e4e8d38c-6293-4531-a8d2-1c8c10848ff4","Type":"ContainerStarted","Data":"c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed"} Dec 03 14:44:23 crc kubenswrapper[4677]: I1203 14:44:23.867008 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zk6bg" podStartSLOduration=3.223684307 podStartE2EDuration="11.866984201s" podCreationTimestamp="2025-12-03 14:44:12 +0000 UTC" firstStartedPulling="2025-12-03 14:44:14.520419109 +0000 UTC m=+3445.266751564" lastFinishedPulling="2025-12-03 14:44:23.163719003 +0000 UTC m=+3453.910051458" observedRunningTime="2025-12-03 14:44:23.809159729 +0000 UTC m=+3454.555492204" watchObservedRunningTime="2025-12-03 14:44:23.866984201 +0000 UTC m=+3454.613316676" Dec 03 14:44:32 crc kubenswrapper[4677]: I1203 14:44:32.850690 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:32 crc kubenswrapper[4677]: I1203 14:44:32.851492 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:32 crc kubenswrapper[4677]: I1203 14:44:32.907317 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:33 crc kubenswrapper[4677]: I1203 14:44:33.954302 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:34 crc kubenswrapper[4677]: I1203 14:44:34.002750 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zk6bg"] Dec 03 14:44:35 crc kubenswrapper[4677]: I1203 14:44:35.933484 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zk6bg" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="registry-server" containerID="cri-o://c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed" gracePeriod=2 Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.460059 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.508428 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-utilities\") pod \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.508533 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-catalog-content\") pod \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.508826 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r8ng\" (UniqueName: \"kubernetes.io/projected/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-kube-api-access-6r8ng\") pod \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\" (UID: \"e4e8d38c-6293-4531-a8d2-1c8c10848ff4\") " Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.511125 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-utilities" (OuterVolumeSpecName: "utilities") pod "e4e8d38c-6293-4531-a8d2-1c8c10848ff4" (UID: "e4e8d38c-6293-4531-a8d2-1c8c10848ff4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.518167 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-kube-api-access-6r8ng" (OuterVolumeSpecName: "kube-api-access-6r8ng") pod "e4e8d38c-6293-4531-a8d2-1c8c10848ff4" (UID: "e4e8d38c-6293-4531-a8d2-1c8c10848ff4"). InnerVolumeSpecName "kube-api-access-6r8ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.611046 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r8ng\" (UniqueName: \"kubernetes.io/projected/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-kube-api-access-6r8ng\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.611454 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.625446 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4e8d38c-6293-4531-a8d2-1c8c10848ff4" (UID: "e4e8d38c-6293-4531-a8d2-1c8c10848ff4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.713356 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e8d38c-6293-4531-a8d2-1c8c10848ff4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.945235 4677 generic.go:334] "Generic (PLEG): container finished" podID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerID="c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed" exitCode=0 Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.945297 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk6bg" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.945685 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk6bg" event={"ID":"e4e8d38c-6293-4531-a8d2-1c8c10848ff4","Type":"ContainerDied","Data":"c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed"} Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.945768 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk6bg" event={"ID":"e4e8d38c-6293-4531-a8d2-1c8c10848ff4","Type":"ContainerDied","Data":"0111782ec6ec9fbc7d5d4346e869ee410c6da6f95a88b6da0c60ce69447daafe"} Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.945793 4677 scope.go:117] "RemoveContainer" containerID="c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.980043 4677 scope.go:117] "RemoveContainer" containerID="b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741" Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.983247 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zk6bg"] Dec 03 14:44:36 crc kubenswrapper[4677]: I1203 14:44:36.995534 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zk6bg"] Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.006446 4677 scope.go:117] "RemoveContainer" containerID="e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a" Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.061551 4677 scope.go:117] "RemoveContainer" containerID="c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed" Dec 03 14:44:37 crc kubenswrapper[4677]: E1203 14:44:37.061944 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed\": container with ID starting with c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed not found: ID does not exist" containerID="c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed" Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.062018 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed"} err="failed to get container status \"c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed\": rpc error: code = NotFound desc = could not find container \"c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed\": container with ID starting with c517c5cfe4aeaae14ff29f0ef8cbc234e542af1ce0366530d90108b75ee62bed not found: ID does not exist" Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.062040 4677 scope.go:117] "RemoveContainer" containerID="b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741" Dec 03 14:44:37 crc kubenswrapper[4677]: E1203 14:44:37.062459 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741\": container with ID starting with b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741 not found: ID does not exist" containerID="b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741" Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.062505 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741"} err="failed to get container status \"b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741\": rpc error: code = NotFound desc = could not find container \"b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741\": container with ID starting with b4186f201cc0e462c5513a5f3acff09db6e0e68b165ae5ddc5fd3baf2aa1e741 not found: ID does not exist" Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.062534 4677 scope.go:117] "RemoveContainer" containerID="e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a" Dec 03 14:44:37 crc kubenswrapper[4677]: E1203 14:44:37.062850 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a\": container with ID starting with e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a not found: ID does not exist" containerID="e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a" Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.062906 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a"} err="failed to get container status \"e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a\": rpc error: code = NotFound desc = could not find container \"e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a\": container with ID starting with e1918e92fd5819c1a9152972dcd86b0c548620e8053435a6e46860ebb0c8e58a not found: ID does not exist" Dec 03 14:44:37 crc kubenswrapper[4677]: I1203 14:44:37.987434 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" path="/var/lib/kubelet/pods/e4e8d38c-6293-4531-a8d2-1c8c10848ff4/volumes" Dec 03 14:44:38 crc kubenswrapper[4677]: I1203 14:44:38.437365 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:44:38 crc kubenswrapper[4677]: I1203 14:44:38.437752 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.150066 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds"] Dec 03 14:45:00 crc kubenswrapper[4677]: E1203 14:45:00.151181 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="extract-content" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.151198 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="extract-content" Dec 03 14:45:00 crc kubenswrapper[4677]: E1203 14:45:00.151219 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="extract-utilities" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.151227 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="extract-utilities" Dec 03 14:45:00 crc kubenswrapper[4677]: E1203 14:45:00.151269 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="registry-server" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.151277 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="registry-server" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.151480 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e8d38c-6293-4531-a8d2-1c8c10848ff4" containerName="registry-server" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.152272 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.155359 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.159651 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.164056 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds"] Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.245894 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzdxr\" (UniqueName: \"kubernetes.io/projected/8213e335-c755-42a3-aa53-f901c7cc24b1-kube-api-access-dzdxr\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.246192 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8213e335-c755-42a3-aa53-f901c7cc24b1-secret-volume\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.246221 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8213e335-c755-42a3-aa53-f901c7cc24b1-config-volume\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.348246 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8213e335-c755-42a3-aa53-f901c7cc24b1-secret-volume\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.348318 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8213e335-c755-42a3-aa53-f901c7cc24b1-config-volume\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.348509 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzdxr\" (UniqueName: \"kubernetes.io/projected/8213e335-c755-42a3-aa53-f901c7cc24b1-kube-api-access-dzdxr\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.349231 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8213e335-c755-42a3-aa53-f901c7cc24b1-config-volume\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.364835 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8213e335-c755-42a3-aa53-f901c7cc24b1-secret-volume\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.367283 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzdxr\" (UniqueName: \"kubernetes.io/projected/8213e335-c755-42a3-aa53-f901c7cc24b1-kube-api-access-dzdxr\") pod \"collect-profiles-29412885-5x9ds\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.481788 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:45:00 crc kubenswrapper[4677]: I1203 14:45:00.956060 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds"] Dec 03 14:45:01 crc kubenswrapper[4677]: I1203 14:45:01.179574 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" event={"ID":"8213e335-c755-42a3-aa53-f901c7cc24b1","Type":"ContainerStarted","Data":"764244aa2d907ea52ca6f44731268ee16955b460d46fd1d31080e5f8fa3420ce"} Dec 03 14:45:08 crc kubenswrapper[4677]: I1203 14:45:08.437039 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:45:08 crc kubenswrapper[4677]: I1203 14:45:08.437614 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:45:09 crc kubenswrapper[4677]: I1203 14:45:09.807475 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 14:45:09 crc kubenswrapper[4677]: I1203 14:45:09.807818 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 03 14:45:11 crc kubenswrapper[4677]: I1203 14:45:11.647093 4677 patch_prober.go:28] interesting pod/controller-manager-6589dc88cb-f294z container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:45:11 crc kubenswrapper[4677]: I1203 14:45:11.647502 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6589dc88cb-f294z" podUID="c803dd49-87e9-424d-9728-c46bd96b36f9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:45:14 crc kubenswrapper[4677]: I1203 14:45:14.807852 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 14:45:17 crc kubenswrapper[4677]: I1203 14:45:17.922348 4677 patch_prober.go:28] interesting pod/console-6d4575bd56-c4z57 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:45:17 crc kubenswrapper[4677]: I1203 14:45:17.922764 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-6d4575bd56-c4z57" podUID="9ad36d4b-b953-479e-ad36-97665e16f672" containerName="console" probeResult="failure" output="Get \"https://10.217.0.53:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:45:18 crc kubenswrapper[4677]: I1203 14:45:18.167175 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-4s5lm" podUID="630fc3ca-d708-4a40-a6a0-dc2047d01769" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:45:19 crc kubenswrapper[4677]: I1203 14:45:19.806350 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 14:45:19 crc kubenswrapper[4677]: I1203 14:45:19.806479 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 03 14:45:19 crc kubenswrapper[4677]: I1203 14:45:19.807666 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"95ba62682528829a689d6d975df64208c2ca8799b7aa66d9f8d77e207beeaaf9"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 03 14:45:19 crc kubenswrapper[4677]: I1203 14:45:19.807799 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-central-agent" containerID="cri-o://95ba62682528829a689d6d975df64208c2ca8799b7aa66d9f8d77e207beeaaf9" gracePeriod=30 Dec 03 14:45:26 crc kubenswrapper[4677]: I1203 14:45:26.582183 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lcgmk" podUID="968d4f67-d25c-48f4-b120-a345dc551e55" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:45:28 crc kubenswrapper[4677]: I1203 14:45:28.168255 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-4s5lm" podUID="630fc3ca-d708-4a40-a6a0-dc2047d01769" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:45:36 crc kubenswrapper[4677]: I1203 14:45:36.580317 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lcgmk" podUID="968d4f67-d25c-48f4-b120-a345dc551e55" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:45:57 crc kubenswrapper[4677]: W1203 14:45:57.203854 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c64.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/user.slice/user-0.slice/session-c64.scope: no such file or directory Dec 03 14:45:57 crc kubenswrapper[4677]: W1203 14:45:57.204840 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c65.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/user.slice/user-0.slice/session-c65.scope: no such file or directory Dec 03 14:45:57 crc kubenswrapper[4677]: W1203 14:45:57.205163 4677 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c66.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/user.slice/user-0.slice/session-c66.scope: no such file or directory Dec 03 14:45:57 crc kubenswrapper[4677]: I1203 14:45:57.234505 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 03 14:45:57 crc kubenswrapper[4677]: I1203 14:45:57.388186 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podUID="882195ac-f224-4563-b00d-0291e40f3204" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.95:8081/healthz\": dial tcp 10.217.0.95:8081: connect: connection refused" Dec 03 14:45:57 crc kubenswrapper[4677]: I1203 14:45:57.388907 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podUID="882195ac-f224-4563-b00d-0291e40f3204" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.95:8081/readyz\": dial tcp 10.217.0.95:8081: connect: connection refused" Dec 03 14:45:57 crc kubenswrapper[4677]: I1203 14:45:57.412851 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:45:57 crc kubenswrapper[4677]: I1203 14:45:57.412892 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:45:57 crc kubenswrapper[4677]: I1203 14:45:57.563171 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46b3c7a220336ed2f24070b4e085e4e629b1f44e1f35f46cb817a51b75e6f587"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:45:57 crc kubenswrapper[4677]: I1203 14:45:57.563528 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://46b3c7a220336ed2f24070b4e085e4e629b1f44e1f35f46cb817a51b75e6f587" gracePeriod=600 Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.147749 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" podUID="226bed84-3bd3-4f3b-ad1a-ab8030a71b34" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/healthz\": dial tcp 10.217.0.78:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.157426 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" podUID="cca481a1-d874-446a-9f63-926247653192" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.82:8081/healthz\": dial tcp 10.217.0.82:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.157548 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" podUID="cca481a1-d874-446a-9f63-926247653192" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.82:8081/readyz\": dial tcp 10.217.0.82:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.158600 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" podUID="5b6f5612-8a33-46f2-8f4e-415d5bc2e807" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.86:8081/healthz\": dial tcp 10.217.0.86:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.158740 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" podUID="5b6f5612-8a33-46f2-8f4e-415d5bc2e807" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.86:8081/readyz\": dial tcp 10.217.0.86:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.176066 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" podUID="226bed84-3bd3-4f3b-ad1a-ab8030a71b34" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/readyz\": dial tcp 10.217.0.78:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.178732 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" podUID="47d9316a-afea-441d-8c42-45e90efe4d5a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/readyz\": dial tcp 10.217.0.91:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.180626 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" podUID="0ffc0c32-9591-4a0d-b7fb-10b8faa85c18" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/healthz\": dial tcp 10.217.0.83:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.180894 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" podUID="0ffc0c32-9591-4a0d-b7fb-10b8faa85c18" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/readyz\": dial tcp 10.217.0.83:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.181927 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" podUID="460e4110-77d5-476e-88d3-d9ccec539f98" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/readyz\": dial tcp 10.217.0.76:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.182128 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" podUID="460e4110-77d5-476e-88d3-d9ccec539f98" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/healthz\": dial tcp 10.217.0.76:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.185183 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" podUID="59757e80-419c-40f9-9ab2-f6dbeff58f7c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/healthz\": dial tcp 10.217.0.88:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.185292 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" podUID="59757e80-419c-40f9-9ab2-f6dbeff58f7c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/readyz\": dial tcp 10.217.0.88:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.185363 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" podUID="8d9537f6-e3da-4acb-af76-4ad3bbc403ab" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.79:8081/healthz\": dial tcp 10.217.0.79:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.185442 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" podUID="8d9537f6-e3da-4acb-af76-4ad3bbc403ab" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.79:8081/readyz\": dial tcp 10.217.0.79:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.192654 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" podUID="1e0b7fb1-cdb4-44a9-9508-4939f1038023" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.85:8081/readyz\": dial tcp 10.217.0.85:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.192769 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" podUID="5236b4d8-b675-4e53-9c7c-c33606436dff" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.77:8081/healthz\": dial tcp 10.217.0.77:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.192856 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" podUID="5236b4d8-b675-4e53-9c7c-c33606436dff" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.77:8081/readyz\": dial tcp 10.217.0.77:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.193839 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" podUID="f5d376fc-835d-44d7-954a-f838a22f7aa8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.97:8081/healthz\": dial tcp 10.217.0.97:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.193928 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" podUID="f5d376fc-835d-44d7-954a-f838a22f7aa8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.97:8081/readyz\": dial tcp 10.217.0.97:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.194499 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podUID="882195ac-f224-4563-b00d-0291e40f3204" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.95:8081/healthz\": dial tcp 10.217.0.95:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.194678 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" podUID="6c72cd7c-f759-4346-a975-4f1cbec9cbf7" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.74:8081/readyz\": dial tcp 10.217.0.74:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.194857 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" podUID="882195ac-f224-4563-b00d-0291e40f3204" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.95:8081/readyz\": dial tcp 10.217.0.95:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.194918 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" podUID="01e63cef-045a-42e0-8776-2f07b1187a40" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.80:8081/readyz\": dial tcp 10.217.0.80:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.194905 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" podUID="de88ce12-43f1-4e18-ad73-f7d2c222c4ca" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/healthz\": dial tcp 10.217.0.89:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.195053 4677 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.237770 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.199562 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" podUID="df1820a3-7030-449d-ad5d-c8bbdb75ad28" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": dial tcp 10.217.0.75:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.200638 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" podUID="7270260d-7534-4b53-be80-69fbee93fdb7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/healthz\": dial tcp 10.217.0.87:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.200705 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" podUID="7270260d-7534-4b53-be80-69fbee93fdb7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/readyz\": dial tcp 10.217.0.87:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.209449 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" podUID="01e63cef-045a-42e0-8776-2f07b1187a40" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.80:8081/healthz\": dial tcp 10.217.0.80:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.210701 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podUID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/readyz\": dial tcp 10.217.0.90:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.214974 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.93:8081/readyz\": dial tcp 10.217.0.93:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.215068 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" podUID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/healthz\": dial tcp 10.217.0.90:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.215100 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" podUID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.93:8081/healthz\": dial tcp 10.217.0.93:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.217487 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" podUID="de88ce12-43f1-4e18-ad73-f7d2c222c4ca" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/readyz\": dial tcp 10.217.0.89:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.198792 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" podUID="df1820a3-7030-449d-ad5d-c8bbdb75ad28" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/healthz\": dial tcp 10.217.0.75:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.259462 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" podUID="1e0b7fb1-cdb4-44a9-9508-4939f1038023" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.85:8081/healthz\": dial tcp 10.217.0.85:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.272892 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" podUID="40fcffeb-3d82-431f-a208-c06bc3f30557" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8081/healthz\": dial tcp 10.217.0.96:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.273298 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" podUID="40fcffeb-3d82-431f-a208-c06bc3f30557" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8081/readyz\": dial tcp 10.217.0.96:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.278754 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" podUID="47d9316a-afea-441d-8c42-45e90efe4d5a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/healthz\": dial tcp 10.217.0.91:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.297631 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" podUID="f6eaef50-3b3e-4fb9-a22a-7db928062fa3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.55:8080/readyz\": EOF" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.316824 4677 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection reset by peer" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.316878 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection reset by peer" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.729203 4677 patch_prober.go:28] interesting pod/console-6d4575bd56-c4z57 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/health\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.729611 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-6d4575bd56-c4z57" podUID="9ad36d4b-b953-479e-ad36-97665e16f672" containerName="console" probeResult="failure" output="Get \"https://10.217.0.53:8443/health\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.832007 4677 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-8qnws container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.832077 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-8qnws" podUID="b693f37f-a924-4f57-a4a5-c9ba03815229" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.901109 4677 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-sxt9n container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.901497 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" podUID="8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.917164 4677 patch_prober.go:28] interesting pod/console-operator-58897d9998-66lfj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.917231 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-66lfj" podUID="01c58bf7-d620-4770-bfc1-b63c29cf32f9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.928129 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.94:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.935021 4677 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-xxds8 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.935091 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xxds8" podUID="d8ce3e58-f4e8-444f-ba23-172619e34134" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.996315 4677 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fq4ql container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.996395 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" podUID="735074c0-8634-44aa-9cfe-cdd1a07633d4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.996455 4677 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fq4ql container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.996471 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fq4ql" podUID="735074c0-8634-44aa-9cfe-cdd1a07633d4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:50.996469 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-4s5lm" podUID="630fc3ca-d708-4a40-a6a0-dc2047d01769" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.023912 4677 generic.go:334] "Generic (PLEG): container finished" podID="aa84d9ae-8578-40d4-a16a-15de08868770" containerID="95ba62682528829a689d6d975df64208c2ca8799b7aa66d9f8d77e207beeaaf9" exitCode=137 Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.024098 4677 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bhbs4 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.024137 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" podUID="e1149c46-0323-453e-a042-4a6e8155364d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.024191 4677 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bhbs4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": context deadline exceeded" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.024207 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bhbs4" podUID="e1149c46-0323-453e-a042-4a6e8155364d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": context deadline exceeded" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.155160 4677 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-sxt9n container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.155207 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sxt9n" podUID="8eb9d7ce-cd02-426c-9f8c-11c4eaef5c41" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.174031 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.174123 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" podUID="af346d56-cdf6-408d-87fc-6431b12a9cd8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.174148 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" podUID="85cba130-9ba2-48b7-917f-14946bc72419" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.94:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.229620 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="46b3c7a220336ed2f24070b4e085e4e629b1f44e1f35f46cb817a51b75e6f587" exitCode=0 Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.774678 4677 patch_prober.go:28] interesting pod/router-default-5444994796-hmrqp container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.775378 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-hmrqp" podUID="8231653d-6197-4f1b-85bf-8f5c260d47fe" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.774937 4677 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-gbcsj container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.775542 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" podUID="2214172a-784b-4626-9e7d-e0a7a0bbfeb7" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.774983 4677 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-gbcsj container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.775684 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gbcsj" podUID="2214172a-784b-4626-9e7d-e0a7a0bbfeb7" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.792854 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-notification-agent" probeResult="failure" output=< Dec 03 14:46:51 crc kubenswrapper[4677]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 03 14:46:51 crc kubenswrapper[4677]: > Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.840664 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-zbfln" podUID="ecdcc3d2-b371-4439-93b8-fe9e6a945ed9" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.71:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.840892 4677 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.660875154s: [/var/lib/containers/storage/overlay/db2d6d3ebbc7bd48c7c2d6d484706ea6a7ec07016fe68a6baece6e7d9be327e3/diff /var/log/pods/openshift-operators_perses-operator-5446b9c989-lm89l_cfed8170-869b-48c2-8b7f-e69466674721/perses-operator/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.857470 4677 trace.go:236] Trace[989902415]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-operators-785mp" (03-Dec-2025 14:46:49.838) (total time: 2019ms): Dec 03 14:46:51 crc kubenswrapper[4677]: Trace[989902415]: [2.019289847s] [2.019289847s] END Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.869115 4677 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.869169 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.880092 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" podUID="6c72cd7c-f759-4346-a975-4f1cbec9cbf7" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.74:8081/healthz\": dial tcp 10.217.0.74:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: E1203 14:46:51.882766 4677 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="54.681s" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.882804 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerDied","Data":"95ba62682528829a689d6d975df64208c2ca8799b7aa66d9f8d77e207beeaaf9"} Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.882839 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.950880 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" podUID="829c01d5-d4d0-40a4-9513-0e60b322d4d0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.81:8081/healthz\": dial tcp 10.217.0.81:8081: connect: connection refused" Dec 03 14:46:51 crc kubenswrapper[4677]: I1203 14:46:51.951060 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" podUID="829c01d5-d4d0-40a4-9513-0e60b322d4d0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.81:8081/readyz\": dial tcp 10.217.0.81:8081: connect: connection refused" Dec 03 14:46:52 crc kubenswrapper[4677]: I1203 14:46:52.106109 4677 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:52 crc kubenswrapper[4677]: I1203 14:46:52.107060 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:52 crc kubenswrapper[4677]: I1203 14:46:52.128329 4677 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.948275663s: [/var/lib/containers/storage/overlay/1234aa9db4a620e77b1dbc1b6dc1f1b224a2eaee1048a570bd7e88bd71a68b5d/diff /var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-ghccd_24ca8470-a377-4746-a521-8af29247ef95/nmstate-metrics/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.349872 4677 patch_prober.go:28] interesting pod/oauth-openshift-9448d8947-cdz7q container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.349921 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-9448d8947-cdz7q" podUID="7ae79228-b546-472d-9a25-9161ecc2f296" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.385338 4677 generic.go:334] "Generic (PLEG): container finished" podID="df1820a3-7030-449d-ad5d-c8bbdb75ad28" containerID="fbab0e8bef13ca695049b6841260cf6e12d72dc2d219d92987fd74355ee598ec" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.397966 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-f8648f98b-pb5w7" podUID="ae71aaf3-8251-4388-ac3a-26121f63a2c0" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.66:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.410896 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-lcgmk" podUID="968d4f67-d25c-48f4-b120-a345dc551e55" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.410916 4677 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-ps6w5 container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.40:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.410989 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" podUID="65806097-c974-427b-a63c-ed84eaa3b039" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.40:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.411018 4677 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-ps6w5 container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.40:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.411046 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-ps6w5" podUID="65806097-c974-427b-a63c-ed84eaa3b039" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.40:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.417032 4677 generic.go:334] "Generic (PLEG): container finished" podID="de88ce12-43f1-4e18-ad73-f7d2c222c4ca" containerID="5bdb7e20ec10b6d3c8af7be02a54a8b50343b31898ee9fe83d955fa68ee6412c" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.441925 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-vrzqb_ef543e1b-8068-4ea3-b32a-61027b32e95d/approver/0.log" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.451207 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef543e1b-8068-4ea3-b32a-61027b32e95d" containerID="ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.501259 4677 generic.go:334] "Generic (PLEG): container finished" podID="cca481a1-d874-446a-9f63-926247653192" containerID="b0b27241762e21a5b5d7d5cb465ba27e03d9e1bf344db5f935746108044e3354" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.553206 4677 generic.go:334] "Generic (PLEG): container finished" podID="da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6" containerID="7dc7fe460118d9e5a9e8caf63764128596e0dbfd850b00313d334fd84ac09b7a" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.588848 4677 generic.go:334] "Generic (PLEG): container finished" podID="7270260d-7534-4b53-be80-69fbee93fdb7" containerID="443180e9f3c2c6acf66542cfa8dea831a10053187d12b4c593c7a91a674ba7db" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.782153 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.866115 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.866177 4677 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f5071c3757b4cdb8e613f663ffd568ada887a6a40fa2ead131e85042a780db91" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.873164 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="072771d7-84c7-4274-8a3d-194f18a09b24" containerName="galera" probeResult="failure" output="command timed out" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.894932 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="a5ae3526-170f-474a-acce-41e5889470c8" containerName="galera" probeResult="failure" output="command timed out" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.895178 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="072771d7-84c7-4274-8a3d-194f18a09b24" containerName="galera" probeResult="failure" output="command timed out" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.923417 4677 generic.go:334] "Generic (PLEG): container finished" podID="8d9537f6-e3da-4acb-af76-4ad3bbc403ab" containerID="f3b1dac967c7a9f50a11ed22efa68324d3cbdb78e484d97c6b2245bdbfa4b40a" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:52.999915 4677 trace.go:236] Trace[696390941]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-x9htr" (03-Dec-2025 14:46:51.742) (total time: 1257ms): Dec 03 14:46:53 crc kubenswrapper[4677]: Trace[696390941]: [1.257828909s] [1.257828909s] END Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.000159 4677 generic.go:334] "Generic (PLEG): container finished" podID="27ee4b91-853f-444b-8663-d913a1d9d1e5" containerID="e32dd67e5f14673801555948e88d4ec16dc84039ae06ae2447aec42ad600fcc4" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.038993 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-ftj4m" podUID="dfeedbc7-41ac-4853-9254-7b8876109abf" containerName="hostpath-provisioner" probeResult="failure" output="Get \"http://10.217.0.41:9898/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.044052 4677 generic.go:334] "Generic (PLEG): container finished" podID="5236b4d8-b675-4e53-9c7c-c33606436dff" containerID="59258f891e0cac6b63d473d3c5c3c85819831bb88da501c3bd3ecd2388e32690" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.159403 4677 generic.go:334] "Generic (PLEG): container finished" podID="40fcffeb-3d82-431f-a208-c06bc3f30557" containerID="e253c6b10b9c6a6c7b28a547eb9d90c3966ddc5e9c7e3b76e9877a8383d0b3fc" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.184217 4677 generic.go:334] "Generic (PLEG): container finished" podID="af346d56-cdf6-408d-87fc-6431b12a9cd8" containerID="28df861f1deec24d0bfa111276d4f8479afacac38364db9c1ad137071747fe55" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.195437 4677 generic.go:334] "Generic (PLEG): container finished" podID="f6eaef50-3b3e-4fb9-a22a-7db928062fa3" containerID="c14473abd96e8132cf374a9890917c6dcd251941767d1768fb28fb26fd6b53cd" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.220500 4677 generic.go:334] "Generic (PLEG): container finished" podID="01e63cef-045a-42e0-8776-2f07b1187a40" containerID="c06768c550e7877666b788cf62dca604e2645de6329bf0401c2f23dda315458c" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.222651 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" podUID="df1820a3-7030-449d-ad5d-c8bbdb75ad28" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": dial tcp 10.217.0.75:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.226006 4677 generic.go:334] "Generic (PLEG): container finished" podID="460e4110-77d5-476e-88d3-d9ccec539f98" containerID="dcd3afe8316ade59fac5f7187529fdf74a93c64f7a9f7da4c8783e7cbaf15982" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.236116 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" podUID="460e4110-77d5-476e-88d3-d9ccec539f98" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/readyz\": dial tcp 10.217.0.76:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.247990 4677 generic.go:334] "Generic (PLEG): container finished" podID="829c01d5-d4d0-40a4-9513-0e60b322d4d0" containerID="591acea476ba57428d50f4c130889d853d795cc53560dc35d8862573a21baa5f" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.251067 4677 generic.go:334] "Generic (PLEG): container finished" podID="59757e80-419c-40f9-9ab2-f6dbeff58f7c" containerID="eb83d118b0281bbb81d291ec63b5e286cc877e329e86b2d622a3155bd5794ba7" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.252657 4677 generic.go:334] "Generic (PLEG): container finished" podID="6c72cd7c-f759-4346-a975-4f1cbec9cbf7" containerID="3b084af4f4ea3e9b7592fedaec83f68517ea892e15d18345595da484a837fa6f" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.258182 4677 generic.go:334] "Generic (PLEG): container finished" podID="0ffc0c32-9591-4a0d-b7fb-10b8faa85c18" containerID="17bed488d40ef0ceeb4f3650e4bffe5bc797b3eeffb42dc55c1985a1cfe4e1dc" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.264324 4677 generic.go:334] "Generic (PLEG): container finished" podID="f5d376fc-835d-44d7-954a-f838a22f7aa8" containerID="1b5fd321cb538290a11e2961fd267282340ef8de13300c944e9c3c5403ef352b" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.270904 4677 generic.go:334] "Generic (PLEG): container finished" podID="5b6f5612-8a33-46f2-8f4e-415d5bc2e807" containerID="dde233158412629ff036ebfb0445934e4eb70f069399d97f2472ae5a5f015886" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.282546 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" podUID="5236b4d8-b675-4e53-9c7c-c33606436dff" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.77:8081/readyz\": dial tcp 10.217.0.77:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.290511 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" podUID="226bed84-3bd3-4f3b-ad1a-ab8030a71b34" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/readyz\": dial tcp 10.217.0.78:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.291534 4677 generic.go:334] "Generic (PLEG): container finished" podID="47d9316a-afea-441d-8c42-45e90efe4d5a" containerID="c8c0dd542f85475071f120804d614a85b0b6e757ac4b9e1dcf43d9e5df871d3f" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.313616 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" podUID="8d9537f6-e3da-4acb-af76-4ad3bbc403ab" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.79:8081/readyz\": dial tcp 10.217.0.79:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.315369 4677 generic.go:334] "Generic (PLEG): container finished" podID="bbb9ba89-c82a-4cb6-8346-a76d2f24fd38" containerID="00e2823b3efa0dc8e6792e579911b5e272d67e9cd14177020b5755ab6d5645aa" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.334434 4677 generic.go:334] "Generic (PLEG): container finished" podID="882195ac-f224-4563-b00d-0291e40f3204" containerID="bb386a05b822f2476f32b947d5b8373fab71333b3ddc12a3ad8cdf3ce00422ac" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.343672 4677 generic.go:334] "Generic (PLEG): container finished" podID="85cba130-9ba2-48b7-917f-14946bc72419" containerID="cb3d9ca474f021a2d3fc8e8873de0a528ffa31dc3a7f0371b27a807a38cae050" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.359752 4677 generic.go:334] "Generic (PLEG): container finished" podID="226bed84-3bd3-4f3b-ad1a-ab8030a71b34" containerID="8a5c872be7aa850a9a5dbc1717b8c0b496a4d8b4785c8ab36f144e9076da670b" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.377276 4677 generic.go:334] "Generic (PLEG): container finished" podID="1e0b7fb1-cdb4-44a9-9508-4939f1038023" containerID="27d73d3d7f548d8436a2cb74a7b1f335d7189963952ef0f4c479a2090575368a" exitCode=1 Dec 03 14:46:53 crc kubenswrapper[4677]: E1203 14:46:53.475230 4677 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.592s" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.475273 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p4d2h"] Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.482659 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.499821 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" podUID="01e63cef-045a-42e0-8776-2f07b1187a40" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.80:8081/readyz\": dial tcp 10.217.0.80:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.548649 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.548687 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"46b3c7a220336ed2f24070b4e085e4e629b1f44e1f35f46cb817a51b75e6f587"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.548713 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" event={"ID":"df1820a3-7030-449d-ad5d-c8bbdb75ad28","Type":"ContainerDied","Data":"fbab0e8bef13ca695049b6841260cf6e12d72dc2d219d92987fd74355ee598ec"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.548730 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" event={"ID":"de88ce12-43f1-4e18-ad73-f7d2c222c4ca","Type":"ContainerDied","Data":"5bdb7e20ec10b6d3c8af7be02a54a8b50343b31898ee9fe83d955fa68ee6412c"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.548745 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerDied","Data":"ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.548763 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" event={"ID":"cca481a1-d874-446a-9f63-926247653192","Type":"ContainerDied","Data":"b0b27241762e21a5b5d7d5cb465ba27e03d9e1bf344db5f935746108044e3354"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.548779 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-td2w5"] Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551559 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-td2w5"] Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551610 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" event={"ID":"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6","Type":"ContainerDied","Data":"7dc7fe460118d9e5a9e8caf63764128596e0dbfd850b00313d334fd84ac09b7a"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551638 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" event={"ID":"7270260d-7534-4b53-be80-69fbee93fdb7","Type":"ContainerDied","Data":"443180e9f3c2c6acf66542cfa8dea831a10053187d12b4c593c7a91a674ba7db"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551657 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f5071c3757b4cdb8e613f663ffd568ada887a6a40fa2ead131e85042a780db91"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551673 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" event={"ID":"8d9537f6-e3da-4acb-af76-4ad3bbc403ab","Type":"ContainerDied","Data":"f3b1dac967c7a9f50a11ed22efa68324d3cbdb78e484d97c6b2245bdbfa4b40a"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551689 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p4d2h"] Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551707 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" event={"ID":"27ee4b91-853f-444b-8663-d913a1d9d1e5","Type":"ContainerDied","Data":"e32dd67e5f14673801555948e88d4ec16dc84039ae06ae2447aec42ad600fcc4"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551721 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" event={"ID":"5236b4d8-b675-4e53-9c7c-c33606436dff","Type":"ContainerDied","Data":"59258f891e0cac6b63d473d3c5c3c85819831bb88da501c3bd3ecd2388e32690"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551736 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" event={"ID":"40fcffeb-3d82-431f-a208-c06bc3f30557","Type":"ContainerDied","Data":"e253c6b10b9c6a6c7b28a547eb9d90c3966ddc5e9c7e3b76e9877a8383d0b3fc"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551755 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" event={"ID":"af346d56-cdf6-408d-87fc-6431b12a9cd8","Type":"ContainerDied","Data":"28df861f1deec24d0bfa111276d4f8479afacac38364db9c1ad137071747fe55"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551770 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" event={"ID":"f6eaef50-3b3e-4fb9-a22a-7db928062fa3","Type":"ContainerDied","Data":"c14473abd96e8132cf374a9890917c6dcd251941767d1768fb28fb26fd6b53cd"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551784 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" event={"ID":"01e63cef-045a-42e0-8776-2f07b1187a40","Type":"ContainerDied","Data":"c06768c550e7877666b788cf62dca604e2645de6329bf0401c2f23dda315458c"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551798 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" event={"ID":"460e4110-77d5-476e-88d3-d9ccec539f98","Type":"ContainerDied","Data":"dcd3afe8316ade59fac5f7187529fdf74a93c64f7a9f7da4c8783e7cbaf15982"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551811 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" event={"ID":"829c01d5-d4d0-40a4-9513-0e60b322d4d0","Type":"ContainerDied","Data":"591acea476ba57428d50f4c130889d853d795cc53560dc35d8862573a21baa5f"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551824 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" event={"ID":"59757e80-419c-40f9-9ab2-f6dbeff58f7c","Type":"ContainerDied","Data":"eb83d118b0281bbb81d291ec63b5e286cc877e329e86b2d622a3155bd5794ba7"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551837 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" event={"ID":"6c72cd7c-f759-4346-a975-4f1cbec9cbf7","Type":"ContainerDied","Data":"3b084af4f4ea3e9b7592fedaec83f68517ea892e15d18345595da484a837fa6f"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551850 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" event={"ID":"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18","Type":"ContainerDied","Data":"17bed488d40ef0ceeb4f3650e4bffe5bc797b3eeffb42dc55c1985a1cfe4e1dc"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551878 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" event={"ID":"8213e335-c755-42a3-aa53-f901c7cc24b1","Type":"ContainerStarted","Data":"6f3585ef9ef7b61ec0bbb1e6b567f9720ddfe2552aa45c5a90419e9fd5657dce"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551891 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" event={"ID":"f5d376fc-835d-44d7-954a-f838a22f7aa8","Type":"ContainerDied","Data":"1b5fd321cb538290a11e2961fd267282340ef8de13300c944e9c3c5403ef352b"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551903 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" event={"ID":"5b6f5612-8a33-46f2-8f4e-415d5bc2e807","Type":"ContainerDied","Data":"dde233158412629ff036ebfb0445934e4eb70f069399d97f2472ae5a5f015886"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551915 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" event={"ID":"47d9316a-afea-441d-8c42-45e90efe4d5a","Type":"ContainerDied","Data":"c8c0dd542f85475071f120804d614a85b0b6e757ac4b9e1dcf43d9e5df871d3f"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551928 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" event={"ID":"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38","Type":"ContainerDied","Data":"00e2823b3efa0dc8e6792e579911b5e272d67e9cd14177020b5755ab6d5645aa"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551941 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" event={"ID":"882195ac-f224-4563-b00d-0291e40f3204","Type":"ContainerDied","Data":"bb386a05b822f2476f32b947d5b8373fab71333b3ddc12a3ad8cdf3ce00422ac"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551977 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" event={"ID":"85cba130-9ba2-48b7-917f-14946bc72419","Type":"ContainerDied","Data":"cb3d9ca474f021a2d3fc8e8873de0a528ffa31dc3a7f0371b27a807a38cae050"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.551991 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" event={"ID":"226bed84-3bd3-4f3b-ad1a-ab8030a71b34","Type":"ContainerDied","Data":"8a5c872be7aa850a9a5dbc1717b8c0b496a4d8b4785c8ab36f144e9076da670b"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.552004 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" event={"ID":"1e0b7fb1-cdb4-44a9-9508-4939f1038023","Type":"ContainerDied","Data":"27d73d3d7f548d8436a2cb74a7b1f335d7189963952ef0f4c479a2090575368a"} Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.553778 4677 scope.go:117] "RemoveContainer" containerID="6416d1efcbe607163b8f2ed30376827b81909e13d167980cc8424a9c11c811b4" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.568664 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.580850 4677 scope.go:117] "RemoveContainer" containerID="59258f891e0cac6b63d473d3c5c3c85819831bb88da501c3bd3ecd2388e32690" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.592371 4677 scope.go:117] "RemoveContainer" containerID="bb386a05b822f2476f32b947d5b8373fab71333b3ddc12a3ad8cdf3ce00422ac" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.600806 4677 scope.go:117] "RemoveContainer" containerID="eb83d118b0281bbb81d291ec63b5e286cc877e329e86b2d622a3155bd5794ba7" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.600920 4677 scope.go:117] "RemoveContainer" containerID="fbab0e8bef13ca695049b6841260cf6e12d72dc2d219d92987fd74355ee598ec" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.604393 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" podUID="829c01d5-d4d0-40a4-9513-0e60b322d4d0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.81:8081/readyz\": dial tcp 10.217.0.81:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.624499 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" podUID="cca481a1-d874-446a-9f63-926247653192" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.82:8081/readyz\": dial tcp 10.217.0.82:8081: connect: connection refused" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.630341 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-utilities\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.630611 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-catalog-content\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.630642 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz2jz\" (UniqueName: \"kubernetes.io/projected/0215e2c8-b813-4223-98d0-ac6fa681d3e0-kube-api-access-zz2jz\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.647788 4677 scope.go:117] "RemoveContainer" containerID="8a5c872be7aa850a9a5dbc1717b8c0b496a4d8b4785c8ab36f144e9076da670b" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.654925 4677 scope.go:117] "RemoveContainer" containerID="17bed488d40ef0ceeb4f3650e4bffe5bc797b3eeffb42dc55c1985a1cfe4e1dc" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.655301 4677 scope.go:117] "RemoveContainer" containerID="ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.659260 4677 scope.go:117] "RemoveContainer" containerID="27d73d3d7f548d8436a2cb74a7b1f335d7189963952ef0f4c479a2090575368a" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.673693 4677 scope.go:117] "RemoveContainer" containerID="3b084af4f4ea3e9b7592fedaec83f68517ea892e15d18345595da484a837fa6f" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.685121 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.685360 4677 scope.go:117] "RemoveContainer" containerID="5bdb7e20ec10b6d3c8af7be02a54a8b50343b31898ee9fe83d955fa68ee6412c" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.705495 4677 scope.go:117] "RemoveContainer" containerID="00e2823b3efa0dc8e6792e579911b5e272d67e9cd14177020b5755ab6d5645aa" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.713998 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.743398 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.745115 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-utilities\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.745206 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-catalog-content\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.745474 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-utilities\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.745735 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-catalog-content\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.745775 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz2jz\" (UniqueName: \"kubernetes.io/projected/0215e2c8-b813-4223-98d0-ac6fa681d3e0-kube-api-access-zz2jz\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.745831 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwxbt\" (UniqueName: \"kubernetes.io/projected/77496c10-3209-439d-9aec-4b1c02eb786d-kube-api-access-lwxbt\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.751591 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-utilities\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.752442 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-catalog-content\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.753606 4677 scope.go:117] "RemoveContainer" containerID="f5071c3757b4cdb8e613f663ffd568ada887a6a40fa2ead131e85042a780db91" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.767677 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.770199 4677 scope.go:117] "RemoveContainer" containerID="f3b1dac967c7a9f50a11ed22efa68324d3cbdb78e484d97c6b2245bdbfa4b40a" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.770253 4677 scope.go:117] "RemoveContainer" containerID="28df861f1deec24d0bfa111276d4f8479afacac38364db9c1ad137071747fe55" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.771325 4677 scope.go:117] "RemoveContainer" containerID="cb3d9ca474f021a2d3fc8e8873de0a528ffa31dc3a7f0371b27a807a38cae050" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.782828 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz2jz\" (UniqueName: \"kubernetes.io/projected/0215e2c8-b813-4223-98d0-ac6fa681d3e0-kube-api-access-zz2jz\") pod \"certified-operators-p4d2h\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.790772 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.799216 4677 scope.go:117] "RemoveContainer" containerID="b0b27241762e21a5b5d7d5cb465ba27e03d9e1bf344db5f935746108044e3354" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.842337 4677 scope.go:117] "RemoveContainer" containerID="c8c0dd542f85475071f120804d614a85b0b6e757ac4b9e1dcf43d9e5df871d3f" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.861879 4677 scope.go:117] "RemoveContainer" containerID="e32dd67e5f14673801555948e88d4ec16dc84039ae06ae2447aec42ad600fcc4" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.862884 4677 scope.go:117] "RemoveContainer" containerID="7dc7fe460118d9e5a9e8caf63764128596e0dbfd850b00313d334fd84ac09b7a" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.878251 4677 scope.go:117] "RemoveContainer" containerID="dde233158412629ff036ebfb0445934e4eb70f069399d97f2472ae5a5f015886" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.879698 4677 scope.go:117] "RemoveContainer" containerID="c14473abd96e8132cf374a9890917c6dcd251941767d1768fb28fb26fd6b53cd" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.880819 4677 scope.go:117] "RemoveContainer" containerID="dcd3afe8316ade59fac5f7187529fdf74a93c64f7a9f7da4c8783e7cbaf15982" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.909285 4677 scope.go:117] "RemoveContainer" containerID="c06768c550e7877666b788cf62dca604e2645de6329bf0401c2f23dda315458c" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.911356 4677 scope.go:117] "RemoveContainer" containerID="e253c6b10b9c6a6c7b28a547eb9d90c3966ddc5e9c7e3b76e9877a8383d0b3fc" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.916207 4677 scope.go:117] "RemoveContainer" containerID="443180e9f3c2c6acf66542cfa8dea831a10053187d12b4c593c7a91a674ba7db" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.928535 4677 scope.go:117] "RemoveContainer" containerID="591acea476ba57428d50f4c130889d853d795cc53560dc35d8862573a21baa5f" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.932140 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.957036 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.957318 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.960037 4677 scope.go:117] "RemoveContainer" containerID="1b5fd321cb538290a11e2961fd267282340ef8de13300c944e9c3c5403ef352b" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.979587 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwxbt\" (UniqueName: \"kubernetes.io/projected/77496c10-3209-439d-9aec-4b1c02eb786d-kube-api-access-lwxbt\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.980036 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-utilities\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:53 crc kubenswrapper[4677]: I1203 14:46:53.982303 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-catalog-content\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.005095 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-catalog-content\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.029472 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-utilities\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.044503 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" podUID="ccc13e6c-e114-482c-8882-fed3fd6324a0" containerName="cert-manager-controller" probeResult="failure" output="Get \"http://10.217.0.70:9403/livez\": dial tcp 10.217.0.70:9403: connect: connection refused" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.065124 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.065163 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.083613 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwxbt\" (UniqueName: \"kubernetes.io/projected/77496c10-3209-439d-9aec-4b1c02eb786d-kube-api-access-lwxbt\") pod \"redhat-marketplace-td2w5\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.088931 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.339908 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.461194 4677 generic.go:334] "Generic (PLEG): container finished" podID="ccc13e6c-e114-482c-8882-fed3fd6324a0" containerID="95df142bc04fb3953490cbcd156763125441ba606c9065241487edcb0e742939" exitCode=1 Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.461334 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" event={"ID":"ccc13e6c-e114-482c-8882-fed3fd6324a0","Type":"ContainerDied","Data":"95df142bc04fb3953490cbcd156763125441ba606c9065241487edcb0e742939"} Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.462025 4677 scope.go:117] "RemoveContainer" containerID="95df142bc04fb3953490cbcd156763125441ba606c9065241487edcb0e742939" Dec 03 14:46:54 crc kubenswrapper[4677]: E1203 14:46:54.514201 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c5c6f36_9c21_487e_85f3_633d81390939.slice/crio-conmon-997363b21bd56001b0537c7a7bd1b795b0dc140421546249babdbcc75146b7a7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40fcffeb_3d82_431f_a208_c06bc3f30557.slice/crio-e253c6b10b9c6a6c7b28a547eb9d90c3966ddc5e9c7e3b76e9877a8383d0b3fc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa84d9ae_8578_40d4_a16a_15de08868770.slice/crio-95ba62682528829a689d6d975df64208c2ca8799b7aa66d9f8d77e207beeaaf9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod226bed84_3bd3_4f3b_ad1a_ab8030a71b34.slice/crio-8a5c872be7aa850a9a5dbc1717b8c0b496a4d8b4785c8ab36f144e9076da670b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf346d56_cdf6_408d_87fc_6431b12a9cd8.slice/crio-28df861f1deec24d0bfa111276d4f8479afacac38364db9c1ad137071747fe55.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ffc0c32_9591_4a0d_b7fb_10b8faa85c18.slice/crio-17bed488d40ef0ceeb4f3650e4bffe5bc797b3eeffb42dc55c1985a1cfe4e1dc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-conmon-ec3eb79f0d2b5095fe61a19157363d6a2cc332c2c81b62d08680b91a6504c398.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85cba130_9ba2_48b7_917f_14946bc72419.slice/crio-cb3d9ca474f021a2d3fc8e8873de0a528ffa31dc3a7f0371b27a807a38cae050.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8213e335_c755_42a3_aa53_f901c7cc24b1.slice/crio-6f3585ef9ef7b61ec0bbb1e6b567f9720ddfe2552aa45c5a90419e9fd5657dce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcca481a1_d874_446a_9f63_926247653192.slice/crio-conmon-b0b27241762e21a5b5d7d5cb465ba27e03d9e1bf344db5f935746108044e3354.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccc13e6c_e114_482c_8882_fed3fd6324a0.slice/crio-95df142bc04fb3953490cbcd156763125441ba606c9065241487edcb0e742939.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e0b7fb1_cdb4_44a9_9508_4939f1038023.slice/crio-27d73d3d7f548d8436a2cb74a7b1f335d7189963952ef0f4c479a2090575368a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8213e335_c755_42a3_aa53_f901c7cc24b1.slice/crio-conmon-6f3585ef9ef7b61ec0bbb1e6b567f9720ddfe2552aa45c5a90419e9fd5657dce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod460e4110_77d5_476e_88d3_d9ccec539f98.slice/crio-dcd3afe8316ade59fac5f7187529fdf74a93c64f7a9f7da4c8783e7cbaf15982.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b6f5612_8a33_46f2_8f4e_415d5bc2e807.slice/crio-dde233158412629ff036ebfb0445934e4eb70f069399d97f2472ae5a5f015886.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf1820a3_7030_449d_ad5d_c8bbdb75ad28.slice/crio-fbab0e8bef13ca695049b6841260cf6e12d72dc2d219d92987fd74355ee598ec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5236b4d8_b675_4e53_9c7c_c33606436dff.slice/crio-conmon-59258f891e0cac6b63d473d3c5c3c85819831bb88da501c3bd3ecd2388e32690.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47d9316a_afea_441d_8c42_45e90efe4d5a.slice/crio-c8c0dd542f85475071f120804d614a85b0b6e757ac4b9e1dcf43d9e5df871d3f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c5c6f36_9c21_487e_85f3_633d81390939.slice/crio-997363b21bd56001b0537c7a7bd1b795b0dc140421546249babdbcc75146b7a7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7270260d_7534_4b53_be80_69fbee93fdb7.slice/crio-443180e9f3c2c6acf66542cfa8dea831a10053187d12b4c593c7a91a674ba7db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf1820a3_7030_449d_ad5d_c8bbdb75ad28.slice/crio-conmon-fbab0e8bef13ca695049b6841260cf6e12d72dc2d219d92987fd74355ee598ec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c72cd7c_f759_4346_a975_4f1cbec9cbf7.slice/crio-3b084af4f4ea3e9b7592fedaec83f68517ea892e15d18345595da484a837fa6f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbb9ba89_c82a_4cb6_8346_a76d2f24fd38.slice/crio-conmon-00e2823b3efa0dc8e6792e579911b5e272d67e9cd14177020b5755ab6d5645aa.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.528929 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-bhtkj_2c5c6f36-9c21-487e-85f3-633d81390939/cluster-samples-operator/0.log" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.529150 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" event={"ID":"2c5c6f36-9c21-487e-85f3-633d81390939","Type":"ContainerDied","Data":"997363b21bd56001b0537c7a7bd1b795b0dc140421546249babdbcc75146b7a7"} Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.529898 4677 scope.go:117] "RemoveContainer" containerID="997363b21bd56001b0537c7a7bd1b795b0dc140421546249babdbcc75146b7a7" Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.529118 4677 generic.go:334] "Generic (PLEG): container finished" podID="2c5c6f36-9c21-487e-85f3-633d81390939" containerID="997363b21bd56001b0537c7a7bd1b795b0dc140421546249babdbcc75146b7a7" exitCode=2 Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.568409 4677 generic.go:334] "Generic (PLEG): container finished" podID="8213e335-c755-42a3-aa53-f901c7cc24b1" containerID="6f3585ef9ef7b61ec0bbb1e6b567f9720ddfe2552aa45c5a90419e9fd5657dce" exitCode=0 Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.568460 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" event={"ID":"8213e335-c755-42a3-aa53-f901c7cc24b1","Type":"ContainerDied","Data":"6f3585ef9ef7b61ec0bbb1e6b567f9720ddfe2552aa45c5a90419e9fd5657dce"} Dec 03 14:46:54 crc kubenswrapper[4677]: I1203 14:46:54.862146 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="072771d7-84c7-4274-8a3d-194f18a09b24" containerName="galera" probeResult="failure" output="command timed out" Dec 03 14:46:55 crc kubenswrapper[4677]: I1203 14:46:55.231402 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:46:55 crc kubenswrapper[4677]: I1203 14:46:55.283214 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:46:55 crc kubenswrapper[4677]: I1203 14:46:55.581289 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" event={"ID":"59757e80-419c-40f9-9ab2-f6dbeff58f7c","Type":"ContainerStarted","Data":"5eeddcf321dcce681288d81c4161deff63da8d24506e7ddb3f2dc6952873baba"} Dec 03 14:46:55 crc kubenswrapper[4677]: I1203 14:46:55.794577 4677 scope.go:117] "RemoveContainer" containerID="42d57424296dbc5f21ee94b52a90792b965b8be4ddd91d7568b7aca1fbce4035" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.447868 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.644873 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.658609 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-vrzqb_ef543e1b-8068-4ea3-b32a-61027b32e95d/approver/0.log" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.660290 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f13ce9ce18060e2ffa594da40a6ad8aaaa3c6f22c77fd9cb51147c8520d9f9b4"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.672608 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" event={"ID":"df1820a3-7030-449d-ad5d-c8bbdb75ad28","Type":"ContainerStarted","Data":"26b113d06c98389ab4b4cea5a8673d86f8b3943dc419e2376cfbe25bde3c2aba"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.673078 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.744451 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" event={"ID":"1e0b7fb1-cdb4-44a9-9508-4939f1038023","Type":"ContainerStarted","Data":"55c280235de570ad4d17592885a64895b5f57656e83c6d307efc4d85412217bf"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.744630 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.835426 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8213e335-c755-42a3-aa53-f901c7cc24b1-config-volume\") pod \"8213e335-c755-42a3-aa53-f901c7cc24b1\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.835746 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzdxr\" (UniqueName: \"kubernetes.io/projected/8213e335-c755-42a3-aa53-f901c7cc24b1-kube-api-access-dzdxr\") pod \"8213e335-c755-42a3-aa53-f901c7cc24b1\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.835788 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8213e335-c755-42a3-aa53-f901c7cc24b1-secret-volume\") pod \"8213e335-c755-42a3-aa53-f901c7cc24b1\" (UID: \"8213e335-c755-42a3-aa53-f901c7cc24b1\") " Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.836330 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" event={"ID":"882195ac-f224-4563-b00d-0291e40f3204","Type":"ContainerStarted","Data":"f0d5f63bcc54649fb1400b128a5735c861fbf472e35a792fc525ecd256bde80a"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.837247 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.838775 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8213e335-c755-42a3-aa53-f901c7cc24b1-config-volume" (OuterVolumeSpecName: "config-volume") pod "8213e335-c755-42a3-aa53-f901c7cc24b1" (UID: "8213e335-c755-42a3-aa53-f901c7cc24b1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.878818 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" event={"ID":"0ffc0c32-9591-4a0d-b7fb-10b8faa85c18","Type":"ContainerStarted","Data":"b303b2f2341470d23ca7eecffc1f41a7d985ff099cf3137c09516ac76915baaa"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.880273 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.931016 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.933913 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" event={"ID":"de88ce12-43f1-4e18-ad73-f7d2c222c4ca","Type":"ContainerStarted","Data":"bf73a92e01450bf9191ab8616fa3162aac23d018848e188a786036a809c9ce81"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.934887 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.937980 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8213e335-c755-42a3-aa53-f901c7cc24b1-kube-api-access-dzdxr" (OuterVolumeSpecName: "kube-api-access-dzdxr") pod "8213e335-c755-42a3-aa53-f901c7cc24b1" (UID: "8213e335-c755-42a3-aa53-f901c7cc24b1"). InnerVolumeSpecName "kube-api-access-dzdxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.943779 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8213e335-c755-42a3-aa53-f901c7cc24b1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8213e335-c755-42a3-aa53-f901c7cc24b1" (UID: "8213e335-c755-42a3-aa53-f901c7cc24b1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.947268 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8213e335-c755-42a3-aa53-f901c7cc24b1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.947313 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzdxr\" (UniqueName: \"kubernetes.io/projected/8213e335-c755-42a3-aa53-f901c7cc24b1-kube-api-access-dzdxr\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.947325 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8213e335-c755-42a3-aa53-f901c7cc24b1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.972496 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" event={"ID":"8213e335-c755-42a3-aa53-f901c7cc24b1","Type":"ContainerDied","Data":"764244aa2d907ea52ca6f44731268ee16955b460d46fd1d31080e5f8fa3420ce"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.972808 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="764244aa2d907ea52ca6f44731268ee16955b460d46fd1d31080e5f8fa3420ce" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.972987 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds" Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.994227 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" event={"ID":"bbb9ba89-c82a-4cb6-8346-a76d2f24fd38","Type":"ContainerStarted","Data":"971b4690137a21dc9cc3afcde6d6468bae07a4086d10c3bc8fc765779bc33297"} Dec 03 14:46:56 crc kubenswrapper[4677]: I1203 14:46:56.997236 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.028281 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" event={"ID":"5236b4d8-b675-4e53-9c7c-c33606436dff","Type":"ContainerStarted","Data":"da9d53fa1447c3aad4b58338a0119f2aedbc257aaf6a8f995ed3389fe1402641"} Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.028838 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.189790 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerStarted","Data":"a30dd375ee1b1f1978173640f8b754b3782ca6a395b9417f7a12e0fb2a46d3eb"} Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.190473 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.191057 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-notification-agent" containerStatusID={"Type":"cri-o","ID":"9d7daf84846edba6140614669fdd4c4b49383f6c9bd32cbad16651d71733b669"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-notification-agent failed liveness probe, will be restarted" Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.191125 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa84d9ae-8578-40d4-a16a-15de08868770" containerName="ceilometer-notification-agent" containerID="cri-o://9d7daf84846edba6140614669fdd4c4b49383f6c9bd32cbad16651d71733b669" gracePeriod=30 Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.447641 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p4d2h"] Dec 03 14:46:57 crc kubenswrapper[4677]: W1203 14:46:57.494895 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0215e2c8_b813_4223_98d0_ac6fa681d3e0.slice/crio-c96994ba0867e751c45506e9b3924125d5692673d246227a2affac2c1d953593 WatchSource:0}: Error finding container c96994ba0867e751c45506e9b3924125d5692673d246227a2affac2c1d953593: Status 404 returned error can't find the container with id c96994ba0867e751c45506e9b3924125d5692673d246227a2affac2c1d953593 Dec 03 14:46:57 crc kubenswrapper[4677]: I1203 14:46:57.764629 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-td2w5"] Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.234448 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4d2h" event={"ID":"0215e2c8-b813-4223-98d0-ac6fa681d3e0","Type":"ContainerStarted","Data":"c96994ba0867e751c45506e9b3924125d5692673d246227a2affac2c1d953593"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.252773 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" event={"ID":"cca481a1-d874-446a-9f63-926247653192","Type":"ContainerStarted","Data":"439f9186bede1befcebda6f4541ee174144b637c32f465fbf51f80ded14e14b9"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.254237 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.290131 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.321822 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" event={"ID":"8d9537f6-e3da-4acb-af76-4ad3bbc403ab","Type":"ContainerStarted","Data":"f65e38ec833c39f56263af7368a91eae8e74d4f95ce32afd69da91b42bb6606f"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.322209 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.328847 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" event={"ID":"460e4110-77d5-476e-88d3-d9ccec539f98","Type":"ContainerStarted","Data":"6e2012fdf9382eb5a193e63dd787fc73553b22adb9dff426d381541540fb8285"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.329330 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.339992 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" event={"ID":"829c01d5-d4d0-40a4-9513-0e60b322d4d0","Type":"ContainerStarted","Data":"590dec59a7237c5c91150538938839a5773e8469d0d8b12f1430acae99db8750"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.340197 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.359183 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" event={"ID":"85cba130-9ba2-48b7-917f-14946bc72419","Type":"ContainerStarted","Data":"00caa00f0d9b067463ba52042aaa88f2e8605015945118908d59f581b8d45330"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.360316 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.376922 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" event={"ID":"226bed84-3bd3-4f3b-ad1a-ab8030a71b34","Type":"ContainerStarted","Data":"d554e502eb1f421feed823ebd464cc146a9d688de500d4d8d4e1af4858cace06"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.378609 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.403762 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-td2w5" event={"ID":"77496c10-3209-439d-9aec-4b1c02eb786d","Type":"ContainerStarted","Data":"a3797657ab11a076cd11245e716cbfa002e9136b73951a99ec3b2731f412fe86"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.413863 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" event={"ID":"6c72cd7c-f759-4346-a975-4f1cbec9cbf7","Type":"ContainerStarted","Data":"6b3c1995d5bf84eabe412021301d4de8b2ee15673bea70015ef0b9007e8d07b1"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.414281 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.457200 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" event={"ID":"af346d56-cdf6-408d-87fc-6431b12a9cd8","Type":"ContainerStarted","Data":"dc4ea6c2b77bce04d8085da85a7b3eba0cbff7067359f2262b208ba2500f356e"} Dec 03 14:46:58 crc kubenswrapper[4677]: I1203 14:46:58.457518 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.346199 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.346733 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.505802 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" event={"ID":"5b6f5612-8a33-46f2-8f4e-415d5bc2e807","Type":"ContainerStarted","Data":"879d05cda7541424e372ac9c9aa5e607636a3445612ca71a2c2324a10f72865f"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.507127 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.519880 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-b5kfm" event={"ID":"ccc13e6c-e114-482c-8882-fed3fd6324a0","Type":"ContainerStarted","Data":"3faf8635e0fcea5133daf2d3de9e02c694ca13617e49231e34045c30b1b8257b"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.546283 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnqg7" event={"ID":"27ee4b91-853f-444b-8663-d913a1d9d1e5","Type":"ContainerStarted","Data":"f1bc450b22540f279073f6c5c53d94488da019fb32ada2a5126f79739b1beb16"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.577812 4677 generic.go:334] "Generic (PLEG): container finished" podID="77496c10-3209-439d-9aec-4b1c02eb786d" containerID="12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2" exitCode=0 Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.577914 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-td2w5" event={"ID":"77496c10-3209-439d-9aec-4b1c02eb786d","Type":"ContainerDied","Data":"12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.583020 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.619354 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" event={"ID":"40fcffeb-3d82-431f-a208-c06bc3f30557","Type":"ContainerStarted","Data":"16655b05730cdb8512c7b99f4889584aa3e7d0513e37067ae5818e0699da692a"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.620508 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.652187 4677 generic.go:334] "Generic (PLEG): container finished" podID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerID="b0bca0307a0e4dd3ad179003d6e077d02cf47c788d10571844a91c5b50690a7b" exitCode=0 Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.652307 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4d2h" event={"ID":"0215e2c8-b813-4223-98d0-ac6fa681d3e0","Type":"ContainerDied","Data":"b0bca0307a0e4dd3ad179003d6e077d02cf47c788d10571844a91c5b50690a7b"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.671324 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" event={"ID":"da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6","Type":"ContainerStarted","Data":"4cd7e9d0c062e9b737a9e58b42d4adb4260fa52a33359b7b0d4a5e11a9a63dfb"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.672267 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.688423 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-bhtkj_2c5c6f36-9c21-487e-85f3-633d81390939/cluster-samples-operator/0.log" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.688541 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhtkj" event={"ID":"2c5c6f36-9c21-487e-85f3-633d81390939","Type":"ContainerStarted","Data":"d86baae8d0095c17dd61de1863138a6e3cdd417c19897307b7410026d879b728"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.716519 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.716659 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" event={"ID":"7270260d-7534-4b53-be80-69fbee93fdb7","Type":"ContainerStarted","Data":"71ba66178172c6bd26380537b434b34772a973f3cae7d5f0d90abadab36fc291"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.716895 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.742432 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.757778 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b93779125159bfbd7e59d18e0241c36f73d186b47a5c1b5713570e0bb4a381f0"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.791255 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.791334 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.792197 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" event={"ID":"f5d376fc-835d-44d7-954a-f838a22f7aa8","Type":"ContainerStarted","Data":"ee235e412f7352bcaa9b9249f044b16206f6727a4179f4c77a17ed1239ed3dc5"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.792563 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.800031 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" event={"ID":"f6eaef50-3b3e-4fb9-a22a-7db928062fa3","Type":"ContainerStarted","Data":"3897c0195c793b41548efe1a1580b68dc25a14b9d49e4d7ce258d1f34999e147"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.801044 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.822741 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" event={"ID":"01e63cef-045a-42e0-8776-2f07b1187a40","Type":"ContainerStarted","Data":"971d003cc4ca3fcf20622c0ad6c1384945f3c40a4b358a8a975c7461d1b3bcbd"} Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.823684 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:46:59 crc kubenswrapper[4677]: I1203 14:46:59.834033 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" event={"ID":"47d9316a-afea-441d-8c42-45e90efe4d5a","Type":"ContainerStarted","Data":"dc2a7ce2357e125fca6beba01e1bef3b978f21d4536c16f7b4d44c19460f9cad"} Dec 03 14:47:00 crc kubenswrapper[4677]: I1203 14:47:00.846864 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-td2w5" event={"ID":"77496c10-3209-439d-9aec-4b1c02eb786d","Type":"ContainerStarted","Data":"d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5"} Dec 03 14:47:00 crc kubenswrapper[4677]: I1203 14:47:00.849569 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4d2h" event={"ID":"0215e2c8-b813-4223-98d0-ac6fa681d3e0","Type":"ContainerStarted","Data":"abb45d281479ba81f4d9710783328e122665ad86094dc7b4d4091f86bdc47265"} Dec 03 14:47:00 crc kubenswrapper[4677]: I1203 14:47:00.854606 4677 generic.go:334] "Generic (PLEG): container finished" podID="aa84d9ae-8578-40d4-a16a-15de08868770" containerID="9d7daf84846edba6140614669fdd4c4b49383f6c9bd32cbad16651d71733b669" exitCode=0 Dec 03 14:47:00 crc kubenswrapper[4677]: I1203 14:47:00.855747 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerDied","Data":"9d7daf84846edba6140614669fdd4c4b49383f6c9bd32cbad16651d71733b669"} Dec 03 14:47:00 crc kubenswrapper[4677]: I1203 14:47:00.855775 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa84d9ae-8578-40d4-a16a-15de08868770","Type":"ContainerStarted","Data":"8a2d8f786e4618070179fbbb273a7eb5fb80a14dc44f286beb79d0da28859805"} Dec 03 14:47:00 crc kubenswrapper[4677]: I1203 14:47:00.860333 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:47:00 crc kubenswrapper[4677]: I1203 14:47:00.894761 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:47:01 crc kubenswrapper[4677]: I1203 14:47:01.871348 4677 generic.go:334] "Generic (PLEG): container finished" podID="77496c10-3209-439d-9aec-4b1c02eb786d" containerID="d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5" exitCode=0 Dec 03 14:47:01 crc kubenswrapper[4677]: I1203 14:47:01.871768 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-td2w5" event={"ID":"77496c10-3209-439d-9aec-4b1c02eb786d","Type":"ContainerDied","Data":"d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5"} Dec 03 14:47:01 crc kubenswrapper[4677]: I1203 14:47:01.877763 4677 generic.go:334] "Generic (PLEG): container finished" podID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerID="abb45d281479ba81f4d9710783328e122665ad86094dc7b4d4091f86bdc47265" exitCode=0 Dec 03 14:47:01 crc kubenswrapper[4677]: I1203 14:47:01.879581 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4d2h" event={"ID":"0215e2c8-b813-4223-98d0-ac6fa681d3e0","Type":"ContainerDied","Data":"abb45d281479ba81f4d9710783328e122665ad86094dc7b4d4091f86bdc47265"} Dec 03 14:47:02 crc kubenswrapper[4677]: I1203 14:47:02.891480 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-td2w5" event={"ID":"77496c10-3209-439d-9aec-4b1c02eb786d","Type":"ContainerStarted","Data":"931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e"} Dec 03 14:47:02 crc kubenswrapper[4677]: I1203 14:47:02.895044 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4d2h" event={"ID":"0215e2c8-b813-4223-98d0-ac6fa681d3e0","Type":"ContainerStarted","Data":"31c2f869cb0c795b8ca133ea931fe8ec48542e8acabcfc56fb9f6da71a21209f"} Dec 03 14:47:02 crc kubenswrapper[4677]: I1203 14:47:02.916274 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-td2w5" podStartSLOduration=66.972185708 podStartE2EDuration="1m9.916253995s" podCreationTimestamp="2025-12-03 14:45:53 +0000 UTC" firstStartedPulling="2025-12-03 14:46:59.582747775 +0000 UTC m=+3610.329080230" lastFinishedPulling="2025-12-03 14:47:02.526816062 +0000 UTC m=+3613.273148517" observedRunningTime="2025-12-03 14:47:02.911217767 +0000 UTC m=+3613.657550232" watchObservedRunningTime="2025-12-03 14:47:02.916253995 +0000 UTC m=+3613.662586450" Dec 03 14:47:02 crc kubenswrapper[4677]: I1203 14:47:02.959287 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p4d2h" podStartSLOduration=69.324890825 podStartE2EDuration="1m11.959267282s" podCreationTimestamp="2025-12-03 14:45:51 +0000 UTC" firstStartedPulling="2025-12-03 14:46:59.661782219 +0000 UTC m=+3610.408114664" lastFinishedPulling="2025-12-03 14:47:02.296158666 +0000 UTC m=+3613.042491121" observedRunningTime="2025-12-03 14:47:02.95372392 +0000 UTC m=+3613.700056365" watchObservedRunningTime="2025-12-03 14:47:02.959267282 +0000 UTC m=+3613.705599737" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.225155 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jxph7" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.232628 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-x2q6z" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.284361 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-45txl" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.288504 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8rll8" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.313577 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vr9nf" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.605364 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-l2wcn" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.606689 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-rcgbn" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.684774 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-n6xjd" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.715120 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5f5nf" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.743617 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-4tw2x" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.769699 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-vk8s9" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.793863 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vv4wp" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.838573 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hfwgd" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.858220 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-qxf8w" Dec 03 14:47:03 crc kubenswrapper[4677]: I1203 14:47:03.942324 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8jjpl" Dec 03 14:47:04 crc kubenswrapper[4677]: I1203 14:47:04.039515 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-jmxdr" Dec 03 14:47:04 crc kubenswrapper[4677]: I1203 14:47:04.059457 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-9jc84" Dec 03 14:47:04 crc kubenswrapper[4677]: I1203 14:47:04.098222 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6757bc7c76-rqljj" Dec 03 14:47:05 crc kubenswrapper[4677]: I1203 14:47:05.234547 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:47:05 crc kubenswrapper[4677]: I1203 14:47:05.234983 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:47:05 crc kubenswrapper[4677]: I1203 14:47:05.283607 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:47:05 crc kubenswrapper[4677]: I1203 14:47:05.283670 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:47:05 crc kubenswrapper[4677]: I1203 14:47:05.294718 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:47:05 crc kubenswrapper[4677]: I1203 14:47:05.333591 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:47:09 crc kubenswrapper[4677]: I1203 14:47:09.352413 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-qhdx7" Dec 03 14:47:09 crc kubenswrapper[4677]: I1203 14:47:09.713216 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:47:09 crc kubenswrapper[4677]: I1203 14:47:09.717120 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:47:09 crc kubenswrapper[4677]: I1203 14:47:09.746626 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk" Dec 03 14:47:09 crc kubenswrapper[4677]: I1203 14:47:09.774416 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7b44f76b7c-m9c5x" Dec 03 14:47:09 crc kubenswrapper[4677]: I1203 14:47:09.798680 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-775cffb9d6-lk7pd" Dec 03 14:47:09 crc kubenswrapper[4677]: I1203 14:47:09.971426 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:47:13 crc kubenswrapper[4677]: I1203 14:47:13.504635 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kqdd2" Dec 03 14:47:15 crc kubenswrapper[4677]: I1203 14:47:15.298978 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:47:15 crc kubenswrapper[4677]: I1203 14:47:15.342391 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:47:15 crc kubenswrapper[4677]: I1203 14:47:15.354764 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p4d2h"] Dec 03 14:47:16 crc kubenswrapper[4677]: I1203 14:47:16.039420 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p4d2h" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="registry-server" containerID="cri-o://31c2f869cb0c795b8ca133ea931fe8ec48542e8acabcfc56fb9f6da71a21209f" gracePeriod=2 Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.077995 4677 generic.go:334] "Generic (PLEG): container finished" podID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerID="31c2f869cb0c795b8ca133ea931fe8ec48542e8acabcfc56fb9f6da71a21209f" exitCode=0 Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.078059 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4d2h" event={"ID":"0215e2c8-b813-4223-98d0-ac6fa681d3e0","Type":"ContainerDied","Data":"31c2f869cb0c795b8ca133ea931fe8ec48542e8acabcfc56fb9f6da71a21209f"} Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.154007 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-td2w5"] Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.154520 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-td2w5" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="registry-server" containerID="cri-o://931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e" gracePeriod=2 Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.533480 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.669609 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz2jz\" (UniqueName: \"kubernetes.io/projected/0215e2c8-b813-4223-98d0-ac6fa681d3e0-kube-api-access-zz2jz\") pod \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.669741 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-utilities\") pod \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.669843 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-catalog-content\") pod \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\" (UID: \"0215e2c8-b813-4223-98d0-ac6fa681d3e0\") " Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.672700 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-utilities" (OuterVolumeSpecName: "utilities") pod "0215e2c8-b813-4223-98d0-ac6fa681d3e0" (UID: "0215e2c8-b813-4223-98d0-ac6fa681d3e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.720891 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0215e2c8-b813-4223-98d0-ac6fa681d3e0-kube-api-access-zz2jz" (OuterVolumeSpecName: "kube-api-access-zz2jz") pod "0215e2c8-b813-4223-98d0-ac6fa681d3e0" (UID: "0215e2c8-b813-4223-98d0-ac6fa681d3e0"). InnerVolumeSpecName "kube-api-access-zz2jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.771853 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz2jz\" (UniqueName: \"kubernetes.io/projected/0215e2c8-b813-4223-98d0-ac6fa681d3e0-kube-api-access-zz2jz\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.771885 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.777645 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0215e2c8-b813-4223-98d0-ac6fa681d3e0" (UID: "0215e2c8-b813-4223-98d0-ac6fa681d3e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:47:17 crc kubenswrapper[4677]: I1203 14:47:17.873745 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0215e2c8-b813-4223-98d0-ac6fa681d3e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.080980 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.096354 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4d2h" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.096372 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4d2h" event={"ID":"0215e2c8-b813-4223-98d0-ac6fa681d3e0","Type":"ContainerDied","Data":"c96994ba0867e751c45506e9b3924125d5692673d246227a2affac2c1d953593"} Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.096440 4677 scope.go:117] "RemoveContainer" containerID="31c2f869cb0c795b8ca133ea931fe8ec48542e8acabcfc56fb9f6da71a21209f" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.101034 4677 generic.go:334] "Generic (PLEG): container finished" podID="77496c10-3209-439d-9aec-4b1c02eb786d" containerID="931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e" exitCode=0 Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.101088 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-td2w5" event={"ID":"77496c10-3209-439d-9aec-4b1c02eb786d","Type":"ContainerDied","Data":"931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e"} Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.101121 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-td2w5" event={"ID":"77496c10-3209-439d-9aec-4b1c02eb786d","Type":"ContainerDied","Data":"a3797657ab11a076cd11245e716cbfa002e9136b73951a99ec3b2731f412fe86"} Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.101200 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-td2w5" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.144281 4677 scope.go:117] "RemoveContainer" containerID="abb45d281479ba81f4d9710783328e122665ad86094dc7b4d4091f86bdc47265" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.161534 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p4d2h"] Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.178942 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-catalog-content\") pod \"77496c10-3209-439d-9aec-4b1c02eb786d\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.179052 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-utilities\") pod \"77496c10-3209-439d-9aec-4b1c02eb786d\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.179097 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwxbt\" (UniqueName: \"kubernetes.io/projected/77496c10-3209-439d-9aec-4b1c02eb786d-kube-api-access-lwxbt\") pod \"77496c10-3209-439d-9aec-4b1c02eb786d\" (UID: \"77496c10-3209-439d-9aec-4b1c02eb786d\") " Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.180880 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-utilities" (OuterVolumeSpecName: "utilities") pod "77496c10-3209-439d-9aec-4b1c02eb786d" (UID: "77496c10-3209-439d-9aec-4b1c02eb786d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.181331 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.178281 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p4d2h"] Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.185980 4677 scope.go:117] "RemoveContainer" containerID="b0bca0307a0e4dd3ad179003d6e077d02cf47c788d10571844a91c5b50690a7b" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.186220 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77496c10-3209-439d-9aec-4b1c02eb786d-kube-api-access-lwxbt" (OuterVolumeSpecName: "kube-api-access-lwxbt") pod "77496c10-3209-439d-9aec-4b1c02eb786d" (UID: "77496c10-3209-439d-9aec-4b1c02eb786d"). InnerVolumeSpecName "kube-api-access-lwxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.202974 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77496c10-3209-439d-9aec-4b1c02eb786d" (UID: "77496c10-3209-439d-9aec-4b1c02eb786d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.237001 4677 scope.go:117] "RemoveContainer" containerID="931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.266824 4677 scope.go:117] "RemoveContainer" containerID="d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.285174 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77496c10-3209-439d-9aec-4b1c02eb786d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.285200 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwxbt\" (UniqueName: \"kubernetes.io/projected/77496c10-3209-439d-9aec-4b1c02eb786d-kube-api-access-lwxbt\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.293105 4677 scope.go:117] "RemoveContainer" containerID="12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.347569 4677 scope.go:117] "RemoveContainer" containerID="931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e" Dec 03 14:47:18 crc kubenswrapper[4677]: E1203 14:47:18.354433 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e\": container with ID starting with 931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e not found: ID does not exist" containerID="931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.354497 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e"} err="failed to get container status \"931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e\": rpc error: code = NotFound desc = could not find container \"931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e\": container with ID starting with 931a870a18364b6ace46e3188b09011eff1820ceeab7f029d9410d4edb73848e not found: ID does not exist" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.354531 4677 scope.go:117] "RemoveContainer" containerID="d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5" Dec 03 14:47:18 crc kubenswrapper[4677]: E1203 14:47:18.355239 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5\": container with ID starting with d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5 not found: ID does not exist" containerID="d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.355277 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5"} err="failed to get container status \"d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5\": rpc error: code = NotFound desc = could not find container \"d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5\": container with ID starting with d8b51797eea640a1d6eacbb96044e269e5ea7aea5c46246f76ab8ae926114ae5 not found: ID does not exist" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.355310 4677 scope.go:117] "RemoveContainer" containerID="12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2" Dec 03 14:47:18 crc kubenswrapper[4677]: E1203 14:47:18.355589 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2\": container with ID starting with 12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2 not found: ID does not exist" containerID="12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.355624 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2"} err="failed to get container status \"12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2\": rpc error: code = NotFound desc = could not find container \"12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2\": container with ID starting with 12b5e3c6c65cbb99716db837589c4af99f0734a6ef3cad3311b654d43e5daaf2 not found: ID does not exist" Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.437376 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-td2w5"] Dec 03 14:47:18 crc kubenswrapper[4677]: I1203 14:47:18.446093 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-td2w5"] Dec 03 14:47:19 crc kubenswrapper[4677]: I1203 14:47:19.989117 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" path="/var/lib/kubelet/pods/0215e2c8-b813-4223-98d0-ac6fa681d3e0/volumes" Dec 03 14:47:19 crc kubenswrapper[4677]: I1203 14:47:19.990675 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" path="/var/lib/kubelet/pods/77496c10-3209-439d-9aec-4b1c02eb786d/volumes" Dec 03 14:47:22 crc kubenswrapper[4677]: I1203 14:47:22.145724 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z"] Dec 03 14:47:22 crc kubenswrapper[4677]: I1203 14:47:22.155412 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-6887z"] Dec 03 14:47:23 crc kubenswrapper[4677]: I1203 14:47:23.991597 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8" path="/var/lib/kubelet/pods/f96388c0-9bb7-4d1f-a8f0-3cd2ad261aa8/volumes" Dec 03 14:47:34 crc kubenswrapper[4677]: I1203 14:47:34.340436 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5f48b5ddf5-58d4r" Dec 03 14:47:50 crc kubenswrapper[4677]: I1203 14:47:50.467936 4677 scope.go:117] "RemoveContainer" containerID="5be8426032319e3c821a9e57f3a73205a4327f696e24458705383cef2a398bac" Dec 03 14:49:08 crc kubenswrapper[4677]: I1203 14:49:08.437222 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:49:08 crc kubenswrapper[4677]: I1203 14:49:08.438110 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:49:38 crc kubenswrapper[4677]: I1203 14:49:38.437080 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:49:38 crc kubenswrapper[4677]: I1203 14:49:38.437680 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:50:08 crc kubenswrapper[4677]: I1203 14:50:08.438314 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:50:08 crc kubenswrapper[4677]: I1203 14:50:08.439082 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:50:08 crc kubenswrapper[4677]: I1203 14:50:08.439143 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:50:08 crc kubenswrapper[4677]: I1203 14:50:08.440037 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:50:08 crc kubenswrapper[4677]: I1203 14:50:08.440103 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" gracePeriod=600 Dec 03 14:50:08 crc kubenswrapper[4677]: E1203 14:50:08.560493 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:50:09 crc kubenswrapper[4677]: I1203 14:50:09.242017 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" exitCode=0 Dec 03 14:50:09 crc kubenswrapper[4677]: I1203 14:50:09.242075 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc"} Dec 03 14:50:09 crc kubenswrapper[4677]: I1203 14:50:09.242145 4677 scope.go:117] "RemoveContainer" containerID="46b3c7a220336ed2f24070b4e085e4e629b1f44e1f35f46cb817a51b75e6f587" Dec 03 14:50:09 crc kubenswrapper[4677]: I1203 14:50:09.242717 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:50:09 crc kubenswrapper[4677]: E1203 14:50:09.243211 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:50:21 crc kubenswrapper[4677]: I1203 14:50:21.977622 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:50:21 crc kubenswrapper[4677]: E1203 14:50:21.978789 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:50:34 crc kubenswrapper[4677]: I1203 14:50:34.976343 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:50:34 crc kubenswrapper[4677]: E1203 14:50:34.977204 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:50:49 crc kubenswrapper[4677]: I1203 14:50:49.994459 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:50:49 crc kubenswrapper[4677]: E1203 14:50:49.995274 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:51:02 crc kubenswrapper[4677]: I1203 14:51:02.976927 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:51:02 crc kubenswrapper[4677]: E1203 14:51:02.977748 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:51:15 crc kubenswrapper[4677]: I1203 14:51:15.976393 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:51:15 crc kubenswrapper[4677]: E1203 14:51:15.977207 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:51:26 crc kubenswrapper[4677]: I1203 14:51:26.975999 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:51:26 crc kubenswrapper[4677]: E1203 14:51:26.976870 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:51:39 crc kubenswrapper[4677]: I1203 14:51:39.986068 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:51:39 crc kubenswrapper[4677]: E1203 14:51:39.987085 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:51:54 crc kubenswrapper[4677]: I1203 14:51:54.976031 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:51:54 crc kubenswrapper[4677]: E1203 14:51:54.976856 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:52:06 crc kubenswrapper[4677]: I1203 14:52:06.977295 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:52:06 crc kubenswrapper[4677]: E1203 14:52:06.978087 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:52:17 crc kubenswrapper[4677]: I1203 14:52:17.977209 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:52:17 crc kubenswrapper[4677]: E1203 14:52:17.978280 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:52:29 crc kubenswrapper[4677]: I1203 14:52:29.984869 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:52:29 crc kubenswrapper[4677]: E1203 14:52:29.985764 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:52:41 crc kubenswrapper[4677]: I1203 14:52:41.976295 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:52:41 crc kubenswrapper[4677]: E1203 14:52:41.977106 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:52:52 crc kubenswrapper[4677]: I1203 14:52:52.976965 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:52:52 crc kubenswrapper[4677]: E1203 14:52:52.979095 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:53:07 crc kubenswrapper[4677]: I1203 14:53:07.976485 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:53:07 crc kubenswrapper[4677]: E1203 14:53:07.978656 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:53:20 crc kubenswrapper[4677]: I1203 14:53:20.975898 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:53:20 crc kubenswrapper[4677]: E1203 14:53:20.977501 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:53:34 crc kubenswrapper[4677]: I1203 14:53:34.977065 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:53:34 crc kubenswrapper[4677]: E1203 14:53:34.978040 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:53:49 crc kubenswrapper[4677]: I1203 14:53:49.985271 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:53:49 crc kubenswrapper[4677]: E1203 14:53:49.986139 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:54:03 crc kubenswrapper[4677]: I1203 14:54:03.975918 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:54:03 crc kubenswrapper[4677]: E1203 14:54:03.976812 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:54:15 crc kubenswrapper[4677]: I1203 14:54:15.976829 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:54:15 crc kubenswrapper[4677]: E1203 14:54:15.977601 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:54:27 crc kubenswrapper[4677]: I1203 14:54:27.977324 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:54:27 crc kubenswrapper[4677]: E1203 14:54:27.978478 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:54:41 crc kubenswrapper[4677]: I1203 14:54:41.976559 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:54:41 crc kubenswrapper[4677]: E1203 14:54:41.977343 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:54:52 crc kubenswrapper[4677]: I1203 14:54:52.976461 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:54:52 crc kubenswrapper[4677]: E1203 14:54:52.977312 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:55:05 crc kubenswrapper[4677]: I1203 14:55:05.977765 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:55:05 crc kubenswrapper[4677]: E1203 14:55:05.978939 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 14:55:17 crc kubenswrapper[4677]: I1203 14:55:17.977326 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:55:18 crc kubenswrapper[4677]: I1203 14:55:18.681625 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"6aaa9b352a55266653b6997926c7a52e8b3205ac500a09ebc2a7d36ecaf352ee"} Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.848418 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w6pg9"] Dec 03 14:56:30 crc kubenswrapper[4677]: E1203 14:56:30.849336 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="extract-utilities" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849353 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="extract-utilities" Dec 03 14:56:30 crc kubenswrapper[4677]: E1203 14:56:30.849372 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="registry-server" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849378 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="registry-server" Dec 03 14:56:30 crc kubenswrapper[4677]: E1203 14:56:30.849402 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8213e335-c755-42a3-aa53-f901c7cc24b1" containerName="collect-profiles" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849408 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="8213e335-c755-42a3-aa53-f901c7cc24b1" containerName="collect-profiles" Dec 03 14:56:30 crc kubenswrapper[4677]: E1203 14:56:30.849420 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="extract-utilities" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849427 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="extract-utilities" Dec 03 14:56:30 crc kubenswrapper[4677]: E1203 14:56:30.849440 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="extract-content" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849446 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="extract-content" Dec 03 14:56:30 crc kubenswrapper[4677]: E1203 14:56:30.849463 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="registry-server" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849469 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="registry-server" Dec 03 14:56:30 crc kubenswrapper[4677]: E1203 14:56:30.849480 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="extract-content" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849487 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="extract-content" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849717 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="8213e335-c755-42a3-aa53-f901c7cc24b1" containerName="collect-profiles" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849733 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="0215e2c8-b813-4223-98d0-ac6fa681d3e0" containerName="registry-server" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.849746 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="77496c10-3209-439d-9aec-4b1c02eb786d" containerName="registry-server" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.851268 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.864366 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6pg9"] Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.876271 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j5q8\" (UniqueName: \"kubernetes.io/projected/28cce37c-5805-4b38-88c2-950cfe750316-kube-api-access-4j5q8\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.876594 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-utilities\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.876708 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-catalog-content\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.979303 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-utilities\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.979458 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-catalog-content\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.979569 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j5q8\" (UniqueName: \"kubernetes.io/projected/28cce37c-5805-4b38-88c2-950cfe750316-kube-api-access-4j5q8\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.980003 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-catalog-content\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:30 crc kubenswrapper[4677]: I1203 14:56:30.980004 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-utilities\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:31 crc kubenswrapper[4677]: I1203 14:56:31.006670 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j5q8\" (UniqueName: \"kubernetes.io/projected/28cce37c-5805-4b38-88c2-950cfe750316-kube-api-access-4j5q8\") pod \"redhat-marketplace-w6pg9\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:31 crc kubenswrapper[4677]: I1203 14:56:31.176920 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:31 crc kubenswrapper[4677]: I1203 14:56:31.697406 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6pg9"] Dec 03 14:56:32 crc kubenswrapper[4677]: I1203 14:56:32.447302 4677 generic.go:334] "Generic (PLEG): container finished" podID="28cce37c-5805-4b38-88c2-950cfe750316" containerID="41c7139ca738b6d8205987961ee15fccbdbd9def38a574bdd1e04f5b67a52fb6" exitCode=0 Dec 03 14:56:32 crc kubenswrapper[4677]: I1203 14:56:32.447407 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6pg9" event={"ID":"28cce37c-5805-4b38-88c2-950cfe750316","Type":"ContainerDied","Data":"41c7139ca738b6d8205987961ee15fccbdbd9def38a574bdd1e04f5b67a52fb6"} Dec 03 14:56:32 crc kubenswrapper[4677]: I1203 14:56:32.447614 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6pg9" event={"ID":"28cce37c-5805-4b38-88c2-950cfe750316","Type":"ContainerStarted","Data":"13aed052c1bedc74b0dfa452e070c3ea87a0fadde52be2028e06f0502f003bea"} Dec 03 14:56:32 crc kubenswrapper[4677]: I1203 14:56:32.449500 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:56:33 crc kubenswrapper[4677]: I1203 14:56:33.460337 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6pg9" event={"ID":"28cce37c-5805-4b38-88c2-950cfe750316","Type":"ContainerStarted","Data":"b49ab81d44ca5cdd7b98fd6eb745d48f6791cb1f23a90cb58ffc716568f45909"} Dec 03 14:56:34 crc kubenswrapper[4677]: I1203 14:56:34.473674 4677 generic.go:334] "Generic (PLEG): container finished" podID="28cce37c-5805-4b38-88c2-950cfe750316" containerID="b49ab81d44ca5cdd7b98fd6eb745d48f6791cb1f23a90cb58ffc716568f45909" exitCode=0 Dec 03 14:56:34 crc kubenswrapper[4677]: I1203 14:56:34.473718 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6pg9" event={"ID":"28cce37c-5805-4b38-88c2-950cfe750316","Type":"ContainerDied","Data":"b49ab81d44ca5cdd7b98fd6eb745d48f6791cb1f23a90cb58ffc716568f45909"} Dec 03 14:56:35 crc kubenswrapper[4677]: I1203 14:56:35.489783 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6pg9" event={"ID":"28cce37c-5805-4b38-88c2-950cfe750316","Type":"ContainerStarted","Data":"74efec07e194ac3563bcfd73acf1757797b7123bcbc7b3223a5cb492228edb0b"} Dec 03 14:56:35 crc kubenswrapper[4677]: I1203 14:56:35.513905 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w6pg9" podStartSLOduration=3.072232667 podStartE2EDuration="5.513888318s" podCreationTimestamp="2025-12-03 14:56:30 +0000 UTC" firstStartedPulling="2025-12-03 14:56:32.449264888 +0000 UTC m=+4183.195597343" lastFinishedPulling="2025-12-03 14:56:34.890920519 +0000 UTC m=+4185.637252994" observedRunningTime="2025-12-03 14:56:35.509402376 +0000 UTC m=+4186.255734851" watchObservedRunningTime="2025-12-03 14:56:35.513888318 +0000 UTC m=+4186.260220773" Dec 03 14:56:37 crc kubenswrapper[4677]: I1203 14:56:37.839834 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pjltv"] Dec 03 14:56:37 crc kubenswrapper[4677]: I1203 14:56:37.842763 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:37 crc kubenswrapper[4677]: I1203 14:56:37.865726 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pjltv"] Dec 03 14:56:37 crc kubenswrapper[4677]: I1203 14:56:37.928909 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvhkf\" (UniqueName: \"kubernetes.io/projected/58595265-683e-4b52-9ba1-853f0f416346-kube-api-access-gvhkf\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:37 crc kubenswrapper[4677]: I1203 14:56:37.929046 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-utilities\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:37 crc kubenswrapper[4677]: I1203 14:56:37.929089 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-catalog-content\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:38 crc kubenswrapper[4677]: I1203 14:56:38.031496 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-utilities\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:38 crc kubenswrapper[4677]: I1203 14:56:38.031600 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-catalog-content\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:38 crc kubenswrapper[4677]: I1203 14:56:38.031811 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvhkf\" (UniqueName: \"kubernetes.io/projected/58595265-683e-4b52-9ba1-853f0f416346-kube-api-access-gvhkf\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:38 crc kubenswrapper[4677]: I1203 14:56:38.031969 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-utilities\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:38 crc kubenswrapper[4677]: I1203 14:56:38.032209 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-catalog-content\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:39 crc kubenswrapper[4677]: I1203 14:56:38.703478 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvhkf\" (UniqueName: \"kubernetes.io/projected/58595265-683e-4b52-9ba1-853f0f416346-kube-api-access-gvhkf\") pod \"certified-operators-pjltv\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:39 crc kubenswrapper[4677]: I1203 14:56:38.771326 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:39 crc kubenswrapper[4677]: I1203 14:56:39.527942 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pjltv"] Dec 03 14:56:39 crc kubenswrapper[4677]: W1203 14:56:39.536748 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58595265_683e_4b52_9ba1_853f0f416346.slice/crio-1eb02b0f92f45056c21c3f06f59a6bec5f9aa3deccc0eef4542a49f8874889a6 WatchSource:0}: Error finding container 1eb02b0f92f45056c21c3f06f59a6bec5f9aa3deccc0eef4542a49f8874889a6: Status 404 returned error can't find the container with id 1eb02b0f92f45056c21c3f06f59a6bec5f9aa3deccc0eef4542a49f8874889a6 Dec 03 14:56:40 crc kubenswrapper[4677]: I1203 14:56:40.546710 4677 generic.go:334] "Generic (PLEG): container finished" podID="58595265-683e-4b52-9ba1-853f0f416346" containerID="3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d" exitCode=0 Dec 03 14:56:40 crc kubenswrapper[4677]: I1203 14:56:40.546782 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjltv" event={"ID":"58595265-683e-4b52-9ba1-853f0f416346","Type":"ContainerDied","Data":"3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d"} Dec 03 14:56:40 crc kubenswrapper[4677]: I1203 14:56:40.546835 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjltv" event={"ID":"58595265-683e-4b52-9ba1-853f0f416346","Type":"ContainerStarted","Data":"1eb02b0f92f45056c21c3f06f59a6bec5f9aa3deccc0eef4542a49f8874889a6"} Dec 03 14:56:41 crc kubenswrapper[4677]: I1203 14:56:41.177126 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:41 crc kubenswrapper[4677]: I1203 14:56:41.177529 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:41 crc kubenswrapper[4677]: I1203 14:56:41.245534 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:41 crc kubenswrapper[4677]: I1203 14:56:41.562771 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjltv" event={"ID":"58595265-683e-4b52-9ba1-853f0f416346","Type":"ContainerStarted","Data":"afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1"} Dec 03 14:56:41 crc kubenswrapper[4677]: I1203 14:56:41.625920 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.576238 4677 generic.go:334] "Generic (PLEG): container finished" podID="58595265-683e-4b52-9ba1-853f0f416346" containerID="afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1" exitCode=0 Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.576307 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjltv" event={"ID":"58595265-683e-4b52-9ba1-853f0f416346","Type":"ContainerDied","Data":"afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1"} Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.659999 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ndtcm"] Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.665752 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.670508 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndtcm"] Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.750760 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvrjz\" (UniqueName: \"kubernetes.io/projected/bd95693e-ee3a-4a1b-96e9-4587811b9b74-kube-api-access-hvrjz\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.751006 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-catalog-content\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.751054 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-utilities\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.852844 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-catalog-content\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.853155 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-utilities\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.853223 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvrjz\" (UniqueName: \"kubernetes.io/projected/bd95693e-ee3a-4a1b-96e9-4587811b9b74-kube-api-access-hvrjz\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.854009 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-utilities\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.854503 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-catalog-content\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:42 crc kubenswrapper[4677]: I1203 14:56:42.876097 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvrjz\" (UniqueName: \"kubernetes.io/projected/bd95693e-ee3a-4a1b-96e9-4587811b9b74-kube-api-access-hvrjz\") pod \"community-operators-ndtcm\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:43 crc kubenswrapper[4677]: I1203 14:56:43.005120 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:43 crc kubenswrapper[4677]: I1203 14:56:43.617824 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndtcm"] Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.419551 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6pg9"] Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.420063 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w6pg9" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="registry-server" containerID="cri-o://74efec07e194ac3563bcfd73acf1757797b7123bcbc7b3223a5cb492228edb0b" gracePeriod=2 Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.600295 4677 generic.go:334] "Generic (PLEG): container finished" podID="28cce37c-5805-4b38-88c2-950cfe750316" containerID="74efec07e194ac3563bcfd73acf1757797b7123bcbc7b3223a5cb492228edb0b" exitCode=0 Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.600362 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6pg9" event={"ID":"28cce37c-5805-4b38-88c2-950cfe750316","Type":"ContainerDied","Data":"74efec07e194ac3563bcfd73acf1757797b7123bcbc7b3223a5cb492228edb0b"} Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.603909 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjltv" event={"ID":"58595265-683e-4b52-9ba1-853f0f416346","Type":"ContainerStarted","Data":"c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec"} Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.605366 4677 generic.go:334] "Generic (PLEG): container finished" podID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerID="7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf" exitCode=0 Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.605411 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndtcm" event={"ID":"bd95693e-ee3a-4a1b-96e9-4587811b9b74","Type":"ContainerDied","Data":"7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf"} Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.605452 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndtcm" event={"ID":"bd95693e-ee3a-4a1b-96e9-4587811b9b74","Type":"ContainerStarted","Data":"d1220608c30fe9b02ab8faba9550a9b296fe1b346914727c3572ad1278e6c6d0"} Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.632859 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pjltv" podStartSLOduration=5.208719258 podStartE2EDuration="7.632839722s" podCreationTimestamp="2025-12-03 14:56:37 +0000 UTC" firstStartedPulling="2025-12-03 14:56:40.548566193 +0000 UTC m=+4191.294898668" lastFinishedPulling="2025-12-03 14:56:42.972686677 +0000 UTC m=+4193.719019132" observedRunningTime="2025-12-03 14:56:44.627059655 +0000 UTC m=+4195.373392110" watchObservedRunningTime="2025-12-03 14:56:44.632839722 +0000 UTC m=+4195.379172177" Dec 03 14:56:44 crc kubenswrapper[4677]: I1203 14:56:44.894113 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.016566 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j5q8\" (UniqueName: \"kubernetes.io/projected/28cce37c-5805-4b38-88c2-950cfe750316-kube-api-access-4j5q8\") pod \"28cce37c-5805-4b38-88c2-950cfe750316\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.016730 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-utilities\") pod \"28cce37c-5805-4b38-88c2-950cfe750316\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.016796 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-catalog-content\") pod \"28cce37c-5805-4b38-88c2-950cfe750316\" (UID: \"28cce37c-5805-4b38-88c2-950cfe750316\") " Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.022908 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28cce37c-5805-4b38-88c2-950cfe750316-kube-api-access-4j5q8" (OuterVolumeSpecName: "kube-api-access-4j5q8") pod "28cce37c-5805-4b38-88c2-950cfe750316" (UID: "28cce37c-5805-4b38-88c2-950cfe750316"). InnerVolumeSpecName "kube-api-access-4j5q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.024844 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-utilities" (OuterVolumeSpecName: "utilities") pod "28cce37c-5805-4b38-88c2-950cfe750316" (UID: "28cce37c-5805-4b38-88c2-950cfe750316"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.029891 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j5q8\" (UniqueName: \"kubernetes.io/projected/28cce37c-5805-4b38-88c2-950cfe750316-kube-api-access-4j5q8\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.029915 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.036879 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28cce37c-5805-4b38-88c2-950cfe750316" (UID: "28cce37c-5805-4b38-88c2-950cfe750316"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.131492 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cce37c-5805-4b38-88c2-950cfe750316-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.620855 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndtcm" event={"ID":"bd95693e-ee3a-4a1b-96e9-4587811b9b74","Type":"ContainerStarted","Data":"2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6"} Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.623841 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6pg9" event={"ID":"28cce37c-5805-4b38-88c2-950cfe750316","Type":"ContainerDied","Data":"13aed052c1bedc74b0dfa452e070c3ea87a0fadde52be2028e06f0502f003bea"} Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.623914 4677 scope.go:117] "RemoveContainer" containerID="74efec07e194ac3563bcfd73acf1757797b7123bcbc7b3223a5cb492228edb0b" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.623867 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6pg9" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.654901 4677 scope.go:117] "RemoveContainer" containerID="b49ab81d44ca5cdd7b98fd6eb745d48f6791cb1f23a90cb58ffc716568f45909" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.671494 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6pg9"] Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.680621 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6pg9"] Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.680889 4677 scope.go:117] "RemoveContainer" containerID="41c7139ca738b6d8205987961ee15fccbdbd9def38a574bdd1e04f5b67a52fb6" Dec 03 14:56:45 crc kubenswrapper[4677]: I1203 14:56:45.993227 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28cce37c-5805-4b38-88c2-950cfe750316" path="/var/lib/kubelet/pods/28cce37c-5805-4b38-88c2-950cfe750316/volumes" Dec 03 14:56:46 crc kubenswrapper[4677]: I1203 14:56:46.634906 4677 generic.go:334] "Generic (PLEG): container finished" podID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerID="2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6" exitCode=0 Dec 03 14:56:46 crc kubenswrapper[4677]: I1203 14:56:46.635001 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndtcm" event={"ID":"bd95693e-ee3a-4a1b-96e9-4587811b9b74","Type":"ContainerDied","Data":"2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6"} Dec 03 14:56:47 crc kubenswrapper[4677]: I1203 14:56:47.649911 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndtcm" event={"ID":"bd95693e-ee3a-4a1b-96e9-4587811b9b74","Type":"ContainerStarted","Data":"c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0"} Dec 03 14:56:47 crc kubenswrapper[4677]: I1203 14:56:47.687699 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ndtcm" podStartSLOduration=3.267377144 podStartE2EDuration="5.687677195s" podCreationTimestamp="2025-12-03 14:56:42 +0000 UTC" firstStartedPulling="2025-12-03 14:56:44.607265407 +0000 UTC m=+4195.353597902" lastFinishedPulling="2025-12-03 14:56:47.027565498 +0000 UTC m=+4197.773897953" observedRunningTime="2025-12-03 14:56:47.674379985 +0000 UTC m=+4198.420712450" watchObservedRunningTime="2025-12-03 14:56:47.687677195 +0000 UTC m=+4198.434009650" Dec 03 14:56:48 crc kubenswrapper[4677]: I1203 14:56:48.771776 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:48 crc kubenswrapper[4677]: I1203 14:56:48.773855 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:48 crc kubenswrapper[4677]: I1203 14:56:48.855166 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:50 crc kubenswrapper[4677]: I1203 14:56:50.148324 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:51 crc kubenswrapper[4677]: I1203 14:56:51.820742 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pjltv"] Dec 03 14:56:52 crc kubenswrapper[4677]: I1203 14:56:52.708819 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pjltv" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="registry-server" containerID="cri-o://c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec" gracePeriod=2 Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.005803 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.006112 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.072436 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.219107 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.348628 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-utilities\") pod \"58595265-683e-4b52-9ba1-853f0f416346\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.348904 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-catalog-content\") pod \"58595265-683e-4b52-9ba1-853f0f416346\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.348952 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvhkf\" (UniqueName: \"kubernetes.io/projected/58595265-683e-4b52-9ba1-853f0f416346-kube-api-access-gvhkf\") pod \"58595265-683e-4b52-9ba1-853f0f416346\" (UID: \"58595265-683e-4b52-9ba1-853f0f416346\") " Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.349599 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-utilities" (OuterVolumeSpecName: "utilities") pod "58595265-683e-4b52-9ba1-853f0f416346" (UID: "58595265-683e-4b52-9ba1-853f0f416346"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.366222 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58595265-683e-4b52-9ba1-853f0f416346-kube-api-access-gvhkf" (OuterVolumeSpecName: "kube-api-access-gvhkf") pod "58595265-683e-4b52-9ba1-853f0f416346" (UID: "58595265-683e-4b52-9ba1-853f0f416346"). InnerVolumeSpecName "kube-api-access-gvhkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.432271 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58595265-683e-4b52-9ba1-853f0f416346" (UID: "58595265-683e-4b52-9ba1-853f0f416346"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.451334 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.451376 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvhkf\" (UniqueName: \"kubernetes.io/projected/58595265-683e-4b52-9ba1-853f0f416346-kube-api-access-gvhkf\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.451392 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58595265-683e-4b52-9ba1-853f0f416346-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.719367 4677 generic.go:334] "Generic (PLEG): container finished" podID="58595265-683e-4b52-9ba1-853f0f416346" containerID="c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec" exitCode=0 Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.719439 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjltv" event={"ID":"58595265-683e-4b52-9ba1-853f0f416346","Type":"ContainerDied","Data":"c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec"} Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.719514 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjltv" event={"ID":"58595265-683e-4b52-9ba1-853f0f416346","Type":"ContainerDied","Data":"1eb02b0f92f45056c21c3f06f59a6bec5f9aa3deccc0eef4542a49f8874889a6"} Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.719544 4677 scope.go:117] "RemoveContainer" containerID="c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.719954 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjltv" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.754341 4677 scope.go:117] "RemoveContainer" containerID="afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.776735 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pjltv"] Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.782845 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.790259 4677 scope.go:117] "RemoveContainer" containerID="3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.791168 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pjltv"] Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.839914 4677 scope.go:117] "RemoveContainer" containerID="c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec" Dec 03 14:56:53 crc kubenswrapper[4677]: E1203 14:56:53.840355 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec\": container with ID starting with c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec not found: ID does not exist" containerID="c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.840444 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec"} err="failed to get container status \"c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec\": rpc error: code = NotFound desc = could not find container \"c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec\": container with ID starting with c019105290ac39b0317cfe16af59deabc546941ccd57d2e165c3a9186dde31ec not found: ID does not exist" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.840514 4677 scope.go:117] "RemoveContainer" containerID="afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1" Dec 03 14:56:53 crc kubenswrapper[4677]: E1203 14:56:53.841041 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1\": container with ID starting with afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1 not found: ID does not exist" containerID="afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.841069 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1"} err="failed to get container status \"afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1\": rpc error: code = NotFound desc = could not find container \"afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1\": container with ID starting with afadc61aacf754c988f75503e25cf0a71808a0d1a22bc987656d72c034379da1 not found: ID does not exist" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.841086 4677 scope.go:117] "RemoveContainer" containerID="3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d" Dec 03 14:56:53 crc kubenswrapper[4677]: E1203 14:56:53.841487 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d\": container with ID starting with 3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d not found: ID does not exist" containerID="3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.841526 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d"} err="failed to get container status \"3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d\": rpc error: code = NotFound desc = could not find container \"3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d\": container with ID starting with 3c1b1450944c132ef39d73cbd64bf98c8343a91a951b4c4edc48509d39ba117d not found: ID does not exist" Dec 03 14:56:53 crc kubenswrapper[4677]: I1203 14:56:53.990537 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58595265-683e-4b52-9ba1-853f0f416346" path="/var/lib/kubelet/pods/58595265-683e-4b52-9ba1-853f0f416346/volumes" Dec 03 14:56:55 crc kubenswrapper[4677]: I1203 14:56:55.024849 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ndtcm"] Dec 03 14:56:55 crc kubenswrapper[4677]: I1203 14:56:55.743027 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ndtcm" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="registry-server" containerID="cri-o://c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0" gracePeriod=2 Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.258186 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.416055 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-catalog-content\") pod \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.416310 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-utilities\") pod \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.416366 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvrjz\" (UniqueName: \"kubernetes.io/projected/bd95693e-ee3a-4a1b-96e9-4587811b9b74-kube-api-access-hvrjz\") pod \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\" (UID: \"bd95693e-ee3a-4a1b-96e9-4587811b9b74\") " Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.417008 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-utilities" (OuterVolumeSpecName: "utilities") pod "bd95693e-ee3a-4a1b-96e9-4587811b9b74" (UID: "bd95693e-ee3a-4a1b-96e9-4587811b9b74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.423710 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd95693e-ee3a-4a1b-96e9-4587811b9b74-kube-api-access-hvrjz" (OuterVolumeSpecName: "kube-api-access-hvrjz") pod "bd95693e-ee3a-4a1b-96e9-4587811b9b74" (UID: "bd95693e-ee3a-4a1b-96e9-4587811b9b74"). InnerVolumeSpecName "kube-api-access-hvrjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.471103 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd95693e-ee3a-4a1b-96e9-4587811b9b74" (UID: "bd95693e-ee3a-4a1b-96e9-4587811b9b74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.524690 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.524732 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvrjz\" (UniqueName: \"kubernetes.io/projected/bd95693e-ee3a-4a1b-96e9-4587811b9b74-kube-api-access-hvrjz\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.524746 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd95693e-ee3a-4a1b-96e9-4587811b9b74-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.757459 4677 generic.go:334] "Generic (PLEG): container finished" podID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerID="c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0" exitCode=0 Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.757509 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndtcm" event={"ID":"bd95693e-ee3a-4a1b-96e9-4587811b9b74","Type":"ContainerDied","Data":"c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0"} Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.757539 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndtcm" event={"ID":"bd95693e-ee3a-4a1b-96e9-4587811b9b74","Type":"ContainerDied","Data":"d1220608c30fe9b02ab8faba9550a9b296fe1b346914727c3572ad1278e6c6d0"} Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.757555 4677 scope.go:117] "RemoveContainer" containerID="c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.757565 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndtcm" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.791421 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ndtcm"] Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.796825 4677 scope.go:117] "RemoveContainer" containerID="2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.802566 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ndtcm"] Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.827584 4677 scope.go:117] "RemoveContainer" containerID="7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.871746 4677 scope.go:117] "RemoveContainer" containerID="c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0" Dec 03 14:56:56 crc kubenswrapper[4677]: E1203 14:56:56.872265 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0\": container with ID starting with c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0 not found: ID does not exist" containerID="c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.872322 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0"} err="failed to get container status \"c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0\": rpc error: code = NotFound desc = could not find container \"c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0\": container with ID starting with c1facaa06c6428afe9a341f9fe54c06cfb536296c73453b254d215d569fa74d0 not found: ID does not exist" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.872355 4677 scope.go:117] "RemoveContainer" containerID="2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6" Dec 03 14:56:56 crc kubenswrapper[4677]: E1203 14:56:56.872759 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6\": container with ID starting with 2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6 not found: ID does not exist" containerID="2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.872802 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6"} err="failed to get container status \"2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6\": rpc error: code = NotFound desc = could not find container \"2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6\": container with ID starting with 2e714700153d947fde99450707f812c28c0b993e29a301fd678745ad30230fc6 not found: ID does not exist" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.872831 4677 scope.go:117] "RemoveContainer" containerID="7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf" Dec 03 14:56:56 crc kubenswrapper[4677]: E1203 14:56:56.873098 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf\": container with ID starting with 7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf not found: ID does not exist" containerID="7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf" Dec 03 14:56:56 crc kubenswrapper[4677]: I1203 14:56:56.873135 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf"} err="failed to get container status \"7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf\": rpc error: code = NotFound desc = could not find container \"7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf\": container with ID starting with 7871f30749cac76f8adc731d5116be8fabcd355fc7d1c860a6ef017c33e930bf not found: ID does not exist" Dec 03 14:56:57 crc kubenswrapper[4677]: I1203 14:56:57.988103 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" path="/var/lib/kubelet/pods/bd95693e-ee3a-4a1b-96e9-4587811b9b74/volumes" Dec 03 14:57:38 crc kubenswrapper[4677]: I1203 14:57:38.437022 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:57:38 crc kubenswrapper[4677]: I1203 14:57:38.437619 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:58:08 crc kubenswrapper[4677]: I1203 14:58:08.437277 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:58:08 crc kubenswrapper[4677]: I1203 14:58:08.437986 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.437037 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.437768 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.437850 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.439218 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6aaa9b352a55266653b6997926c7a52e8b3205ac500a09ebc2a7d36ecaf352ee"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.439330 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://6aaa9b352a55266653b6997926c7a52e8b3205ac500a09ebc2a7d36ecaf352ee" gracePeriod=600 Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.855392 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="6aaa9b352a55266653b6997926c7a52e8b3205ac500a09ebc2a7d36ecaf352ee" exitCode=0 Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.855629 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"6aaa9b352a55266653b6997926c7a52e8b3205ac500a09ebc2a7d36ecaf352ee"} Dec 03 14:58:38 crc kubenswrapper[4677]: I1203 14:58:38.855833 4677 scope.go:117] "RemoveContainer" containerID="9cb99fb7196b2da523ae7d20eb84941e4c7c1270d52c0e597da2b53034715dcc" Dec 03 14:58:39 crc kubenswrapper[4677]: I1203 14:58:39.866116 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d"} Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.163640 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x"] Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165370 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="extract-utilities" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165398 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="extract-utilities" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165422 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165431 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165443 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="extract-content" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165451 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="extract-content" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165481 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165489 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165509 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="extract-content" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165516 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="extract-content" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165545 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="extract-utilities" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165554 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="extract-utilities" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165567 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="extract-content" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165574 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="extract-content" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165589 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165597 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: E1203 15:00:00.165614 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="extract-utilities" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165623 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="extract-utilities" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165900 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd95693e-ee3a-4a1b-96e9-4587811b9b74" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165933 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="58595265-683e-4b52-9ba1-853f0f416346" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.165974 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="28cce37c-5805-4b38-88c2-950cfe750316" containerName="registry-server" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.166908 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.169517 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.169609 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.175295 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x"] Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.273903 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7b10ca4-5bf0-44c8-a83a-856deebf4170-secret-volume\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.274649 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7b10ca4-5bf0-44c8-a83a-856deebf4170-config-volume\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.275341 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvnns\" (UniqueName: \"kubernetes.io/projected/a7b10ca4-5bf0-44c8-a83a-856deebf4170-kube-api-access-pvnns\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.377556 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvnns\" (UniqueName: \"kubernetes.io/projected/a7b10ca4-5bf0-44c8-a83a-856deebf4170-kube-api-access-pvnns\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.377657 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7b10ca4-5bf0-44c8-a83a-856deebf4170-secret-volume\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.377743 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7b10ca4-5bf0-44c8-a83a-856deebf4170-config-volume\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.378604 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7b10ca4-5bf0-44c8-a83a-856deebf4170-config-volume\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.387122 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7b10ca4-5bf0-44c8-a83a-856deebf4170-secret-volume\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.396198 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvnns\" (UniqueName: \"kubernetes.io/projected/a7b10ca4-5bf0-44c8-a83a-856deebf4170-kube-api-access-pvnns\") pod \"collect-profiles-29412900-sbr7x\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:00 crc kubenswrapper[4677]: I1203 15:00:00.507154 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:01 crc kubenswrapper[4677]: I1203 15:00:01.023291 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x"] Dec 03 15:00:01 crc kubenswrapper[4677]: I1203 15:00:01.843284 4677 generic.go:334] "Generic (PLEG): container finished" podID="a7b10ca4-5bf0-44c8-a83a-856deebf4170" containerID="c189a7f5fa5d0352318dee22279406cf8e8c9198cb2c292acb9ab0a30b42031d" exitCode=0 Dec 03 15:00:01 crc kubenswrapper[4677]: I1203 15:00:01.843339 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" event={"ID":"a7b10ca4-5bf0-44c8-a83a-856deebf4170","Type":"ContainerDied","Data":"c189a7f5fa5d0352318dee22279406cf8e8c9198cb2c292acb9ab0a30b42031d"} Dec 03 15:00:01 crc kubenswrapper[4677]: I1203 15:00:01.843376 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" event={"ID":"a7b10ca4-5bf0-44c8-a83a-856deebf4170","Type":"ContainerStarted","Data":"e3b98a9322993dead3559cbd37c65ccda458b8e3ec59ff0896a4638c3ed4e044"} Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.258536 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.442756 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7b10ca4-5bf0-44c8-a83a-856deebf4170-config-volume\") pod \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.442871 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7b10ca4-5bf0-44c8-a83a-856deebf4170-secret-volume\") pod \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.442923 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvnns\" (UniqueName: \"kubernetes.io/projected/a7b10ca4-5bf0-44c8-a83a-856deebf4170-kube-api-access-pvnns\") pod \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\" (UID: \"a7b10ca4-5bf0-44c8-a83a-856deebf4170\") " Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.443840 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7b10ca4-5bf0-44c8-a83a-856deebf4170-config-volume" (OuterVolumeSpecName: "config-volume") pod "a7b10ca4-5bf0-44c8-a83a-856deebf4170" (UID: "a7b10ca4-5bf0-44c8-a83a-856deebf4170"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.449609 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b10ca4-5bf0-44c8-a83a-856deebf4170-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a7b10ca4-5bf0-44c8-a83a-856deebf4170" (UID: "a7b10ca4-5bf0-44c8-a83a-856deebf4170"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.449795 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b10ca4-5bf0-44c8-a83a-856deebf4170-kube-api-access-pvnns" (OuterVolumeSpecName: "kube-api-access-pvnns") pod "a7b10ca4-5bf0-44c8-a83a-856deebf4170" (UID: "a7b10ca4-5bf0-44c8-a83a-856deebf4170"). InnerVolumeSpecName "kube-api-access-pvnns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.545906 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7b10ca4-5bf0-44c8-a83a-856deebf4170-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.545940 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7b10ca4-5bf0-44c8-a83a-856deebf4170-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.545973 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvnns\" (UniqueName: \"kubernetes.io/projected/a7b10ca4-5bf0-44c8-a83a-856deebf4170-kube-api-access-pvnns\") on node \"crc\" DevicePath \"\"" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.865072 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" event={"ID":"a7b10ca4-5bf0-44c8-a83a-856deebf4170","Type":"ContainerDied","Data":"e3b98a9322993dead3559cbd37c65ccda458b8e3ec59ff0896a4638c3ed4e044"} Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.865147 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3b98a9322993dead3559cbd37c65ccda458b8e3ec59ff0896a4638c3ed4e044" Dec 03 15:00:03 crc kubenswrapper[4677]: I1203 15:00:03.865206 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x" Dec 03 15:00:04 crc kubenswrapper[4677]: I1203 15:00:04.334252 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt"] Dec 03 15:00:04 crc kubenswrapper[4677]: I1203 15:00:04.346066 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-s44zt"] Dec 03 15:00:05 crc kubenswrapper[4677]: I1203 15:00:05.991352 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a4aeaa-b2e5-4d62-9c8d-be9774befcaf" path="/var/lib/kubelet/pods/94a4aeaa-b2e5-4d62-9c8d-be9774befcaf/volumes" Dec 03 15:00:38 crc kubenswrapper[4677]: I1203 15:00:38.437125 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:00:38 crc kubenswrapper[4677]: I1203 15:00:38.437902 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:00:44 crc kubenswrapper[4677]: I1203 15:00:44.861075 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dvffg"] Dec 03 15:00:44 crc kubenswrapper[4677]: E1203 15:00:44.862378 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b10ca4-5bf0-44c8-a83a-856deebf4170" containerName="collect-profiles" Dec 03 15:00:44 crc kubenswrapper[4677]: I1203 15:00:44.862395 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b10ca4-5bf0-44c8-a83a-856deebf4170" containerName="collect-profiles" Dec 03 15:00:44 crc kubenswrapper[4677]: I1203 15:00:44.862613 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b10ca4-5bf0-44c8-a83a-856deebf4170" containerName="collect-profiles" Dec 03 15:00:44 crc kubenswrapper[4677]: I1203 15:00:44.864263 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:44 crc kubenswrapper[4677]: I1203 15:00:44.875403 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvffg"] Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.048354 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsk5q\" (UniqueName: \"kubernetes.io/projected/b557bbbe-e8bc-493c-9b73-713665a1ca2b-kube-api-access-rsk5q\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.048572 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-catalog-content\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.048613 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-utilities\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.150128 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-catalog-content\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.150176 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-utilities\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.150282 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsk5q\" (UniqueName: \"kubernetes.io/projected/b557bbbe-e8bc-493c-9b73-713665a1ca2b-kube-api-access-rsk5q\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.151651 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-catalog-content\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.152067 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-utilities\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.174821 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsk5q\" (UniqueName: \"kubernetes.io/projected/b557bbbe-e8bc-493c-9b73-713665a1ca2b-kube-api-access-rsk5q\") pod \"redhat-operators-dvffg\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:45 crc kubenswrapper[4677]: I1203 15:00:45.187824 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:46 crc kubenswrapper[4677]: I1203 15:00:45.789064 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvffg"] Dec 03 15:00:46 crc kubenswrapper[4677]: I1203 15:00:46.368194 4677 generic.go:334] "Generic (PLEG): container finished" podID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerID="2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75" exitCode=0 Dec 03 15:00:46 crc kubenswrapper[4677]: I1203 15:00:46.368298 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvffg" event={"ID":"b557bbbe-e8bc-493c-9b73-713665a1ca2b","Type":"ContainerDied","Data":"2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75"} Dec 03 15:00:46 crc kubenswrapper[4677]: I1203 15:00:46.368440 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvffg" event={"ID":"b557bbbe-e8bc-493c-9b73-713665a1ca2b","Type":"ContainerStarted","Data":"ee43378f19fc91f0d98ce402a8cee9e3eed49a7b18d160063ae472f4773a2393"} Dec 03 15:00:47 crc kubenswrapper[4677]: I1203 15:00:47.381172 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvffg" event={"ID":"b557bbbe-e8bc-493c-9b73-713665a1ca2b","Type":"ContainerStarted","Data":"e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a"} Dec 03 15:00:50 crc kubenswrapper[4677]: I1203 15:00:50.982918 4677 scope.go:117] "RemoveContainer" containerID="aeb7d313a45f0f2de4ec4a6bd4786ab579ca0bee5aa3482accfd3b608a2445b6" Dec 03 15:00:51 crc kubenswrapper[4677]: I1203 15:00:51.421706 4677 generic.go:334] "Generic (PLEG): container finished" podID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerID="e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a" exitCode=0 Dec 03 15:00:51 crc kubenswrapper[4677]: I1203 15:00:51.421793 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvffg" event={"ID":"b557bbbe-e8bc-493c-9b73-713665a1ca2b","Type":"ContainerDied","Data":"e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a"} Dec 03 15:00:52 crc kubenswrapper[4677]: I1203 15:00:52.431556 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvffg" event={"ID":"b557bbbe-e8bc-493c-9b73-713665a1ca2b","Type":"ContainerStarted","Data":"43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3"} Dec 03 15:00:52 crc kubenswrapper[4677]: I1203 15:00:52.470330 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dvffg" podStartSLOduration=2.996992746 podStartE2EDuration="8.470307102s" podCreationTimestamp="2025-12-03 15:00:44 +0000 UTC" firstStartedPulling="2025-12-03 15:00:46.369921731 +0000 UTC m=+4437.116254176" lastFinishedPulling="2025-12-03 15:00:51.843236077 +0000 UTC m=+4442.589568532" observedRunningTime="2025-12-03 15:00:52.461939013 +0000 UTC m=+4443.208271488" watchObservedRunningTime="2025-12-03 15:00:52.470307102 +0000 UTC m=+4443.216639557" Dec 03 15:00:55 crc kubenswrapper[4677]: I1203 15:00:55.188684 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:55 crc kubenswrapper[4677]: I1203 15:00:55.188967 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:00:56 crc kubenswrapper[4677]: I1203 15:00:56.266223 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dvffg" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="registry-server" probeResult="failure" output=< Dec 03 15:00:56 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 15:00:56 crc kubenswrapper[4677]: > Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.168674 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412901-55bb9"] Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.172193 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.198411 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412901-55bb9"] Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.275282 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-config-data\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.275344 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-combined-ca-bundle\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.275372 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4j74\" (UniqueName: \"kubernetes.io/projected/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-kube-api-access-s4j74\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.275398 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-fernet-keys\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.376641 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-combined-ca-bundle\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.376677 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4j74\" (UniqueName: \"kubernetes.io/projected/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-kube-api-access-s4j74\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.376706 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-fernet-keys\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.376860 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-config-data\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.383644 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-fernet-keys\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.384190 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-config-data\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.386211 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-combined-ca-bundle\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.408785 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4j74\" (UniqueName: \"kubernetes.io/projected/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-kube-api-access-s4j74\") pod \"keystone-cron-29412901-55bb9\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.497180 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:00 crc kubenswrapper[4677]: W1203 15:01:00.988577 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda2e0a2c_7e17_4c0c_8cee_a2e2cee92632.slice/crio-de9b4598d94314675c1305d0a402af442d0c1823295d9f2fd8b27efe9fe46480 WatchSource:0}: Error finding container de9b4598d94314675c1305d0a402af442d0c1823295d9f2fd8b27efe9fe46480: Status 404 returned error can't find the container with id de9b4598d94314675c1305d0a402af442d0c1823295d9f2fd8b27efe9fe46480 Dec 03 15:01:00 crc kubenswrapper[4677]: I1203 15:01:00.992257 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412901-55bb9"] Dec 03 15:01:01 crc kubenswrapper[4677]: I1203 15:01:01.532299 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-55bb9" event={"ID":"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632","Type":"ContainerStarted","Data":"c6cdbb053c8b3a63f2a814cc4f693185d334917441e7462102027c8de203061d"} Dec 03 15:01:01 crc kubenswrapper[4677]: I1203 15:01:01.532624 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-55bb9" event={"ID":"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632","Type":"ContainerStarted","Data":"de9b4598d94314675c1305d0a402af442d0c1823295d9f2fd8b27efe9fe46480"} Dec 03 15:01:01 crc kubenswrapper[4677]: I1203 15:01:01.570350 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29412901-55bb9" podStartSLOduration=1.5703239359999999 podStartE2EDuration="1.570323936s" podCreationTimestamp="2025-12-03 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:01:01.549104764 +0000 UTC m=+4452.295437279" watchObservedRunningTime="2025-12-03 15:01:01.570323936 +0000 UTC m=+4452.316656421" Dec 03 15:01:05 crc kubenswrapper[4677]: I1203 15:01:05.246344 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:01:05 crc kubenswrapper[4677]: I1203 15:01:05.304559 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:01:05 crc kubenswrapper[4677]: I1203 15:01:05.480812 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvffg"] Dec 03 15:01:05 crc kubenswrapper[4677]: I1203 15:01:05.579059 4677 generic.go:334] "Generic (PLEG): container finished" podID="da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" containerID="c6cdbb053c8b3a63f2a814cc4f693185d334917441e7462102027c8de203061d" exitCode=0 Dec 03 15:01:05 crc kubenswrapper[4677]: I1203 15:01:05.579224 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-55bb9" event={"ID":"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632","Type":"ContainerDied","Data":"c6cdbb053c8b3a63f2a814cc4f693185d334917441e7462102027c8de203061d"} Dec 03 15:01:06 crc kubenswrapper[4677]: I1203 15:01:06.604522 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dvffg" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="registry-server" containerID="cri-o://43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3" gracePeriod=2 Dec 03 15:01:06 crc kubenswrapper[4677]: I1203 15:01:06.964776 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.131925 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-config-data\") pod \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.131995 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4j74\" (UniqueName: \"kubernetes.io/projected/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-kube-api-access-s4j74\") pod \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.134098 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-combined-ca-bundle\") pod \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.134454 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-fernet-keys\") pod \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\" (UID: \"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632\") " Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.141030 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-kube-api-access-s4j74" (OuterVolumeSpecName: "kube-api-access-s4j74") pod "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" (UID: "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632"). InnerVolumeSpecName "kube-api-access-s4j74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.141666 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" (UID: "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.175569 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.197000 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" (UID: "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.234104 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-config-data" (OuterVolumeSpecName: "config-data") pod "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" (UID: "da2e0a2c-7e17-4c0c-8cee-a2e2cee92632"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.237900 4677 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.238069 4677 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.238119 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4j74\" (UniqueName: \"kubernetes.io/projected/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-kube-api-access-s4j74\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.238140 4677 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da2e0a2c-7e17-4c0c-8cee-a2e2cee92632-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.339216 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsk5q\" (UniqueName: \"kubernetes.io/projected/b557bbbe-e8bc-493c-9b73-713665a1ca2b-kube-api-access-rsk5q\") pod \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.339371 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-catalog-content\") pod \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.339524 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-utilities\") pod \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\" (UID: \"b557bbbe-e8bc-493c-9b73-713665a1ca2b\") " Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.340348 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-utilities" (OuterVolumeSpecName: "utilities") pod "b557bbbe-e8bc-493c-9b73-713665a1ca2b" (UID: "b557bbbe-e8bc-493c-9b73-713665a1ca2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.343625 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b557bbbe-e8bc-493c-9b73-713665a1ca2b-kube-api-access-rsk5q" (OuterVolumeSpecName: "kube-api-access-rsk5q") pod "b557bbbe-e8bc-493c-9b73-713665a1ca2b" (UID: "b557bbbe-e8bc-493c-9b73-713665a1ca2b"). InnerVolumeSpecName "kube-api-access-rsk5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.442838 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.442880 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsk5q\" (UniqueName: \"kubernetes.io/projected/b557bbbe-e8bc-493c-9b73-713665a1ca2b-kube-api-access-rsk5q\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.480699 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b557bbbe-e8bc-493c-9b73-713665a1ca2b" (UID: "b557bbbe-e8bc-493c-9b73-713665a1ca2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.544567 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b557bbbe-e8bc-493c-9b73-713665a1ca2b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.613688 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-55bb9" event={"ID":"da2e0a2c-7e17-4c0c-8cee-a2e2cee92632","Type":"ContainerDied","Data":"de9b4598d94314675c1305d0a402af442d0c1823295d9f2fd8b27efe9fe46480"} Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.614549 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de9b4598d94314675c1305d0a402af442d0c1823295d9f2fd8b27efe9fe46480" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.613728 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-55bb9" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.619758 4677 generic.go:334] "Generic (PLEG): container finished" podID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerID="43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3" exitCode=0 Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.619830 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvffg" event={"ID":"b557bbbe-e8bc-493c-9b73-713665a1ca2b","Type":"ContainerDied","Data":"43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3"} Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.619874 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvffg" event={"ID":"b557bbbe-e8bc-493c-9b73-713665a1ca2b","Type":"ContainerDied","Data":"ee43378f19fc91f0d98ce402a8cee9e3eed49a7b18d160063ae472f4773a2393"} Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.619909 4677 scope.go:117] "RemoveContainer" containerID="43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.620168 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvffg" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.649924 4677 scope.go:117] "RemoveContainer" containerID="e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.678885 4677 scope.go:117] "RemoveContainer" containerID="2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.679070 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvffg"] Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.689291 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dvffg"] Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.697126 4677 scope.go:117] "RemoveContainer" containerID="43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3" Dec 03 15:01:07 crc kubenswrapper[4677]: E1203 15:01:07.697590 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3\": container with ID starting with 43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3 not found: ID does not exist" containerID="43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.697620 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3"} err="failed to get container status \"43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3\": rpc error: code = NotFound desc = could not find container \"43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3\": container with ID starting with 43a9bef394aa46a9383e086a33eaaed8c0113b322176ad5753819255e1cb7ec3 not found: ID does not exist" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.697640 4677 scope.go:117] "RemoveContainer" containerID="e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a" Dec 03 15:01:07 crc kubenswrapper[4677]: E1203 15:01:07.698143 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a\": container with ID starting with e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a not found: ID does not exist" containerID="e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.698200 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a"} err="failed to get container status \"e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a\": rpc error: code = NotFound desc = could not find container \"e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a\": container with ID starting with e4bedb5bbfaba8585cb3b5f0377682bbaebe8e52b13f4d372323cf69f4d1955a not found: ID does not exist" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.698238 4677 scope.go:117] "RemoveContainer" containerID="2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75" Dec 03 15:01:07 crc kubenswrapper[4677]: E1203 15:01:07.698511 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75\": container with ID starting with 2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75 not found: ID does not exist" containerID="2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.698543 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75"} err="failed to get container status \"2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75\": rpc error: code = NotFound desc = could not find container \"2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75\": container with ID starting with 2b658a5c44e133efd19189d52b6f05b8fed7e47052b00ff01be4647ca1209c75 not found: ID does not exist" Dec 03 15:01:07 crc kubenswrapper[4677]: I1203 15:01:07.989461 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" path="/var/lib/kubelet/pods/b557bbbe-e8bc-493c-9b73-713665a1ca2b/volumes" Dec 03 15:01:08 crc kubenswrapper[4677]: I1203 15:01:08.437851 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:01:08 crc kubenswrapper[4677]: I1203 15:01:08.438307 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:01:38 crc kubenswrapper[4677]: I1203 15:01:38.437918 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:01:38 crc kubenswrapper[4677]: I1203 15:01:38.438832 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:01:38 crc kubenswrapper[4677]: I1203 15:01:38.438908 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:01:38 crc kubenswrapper[4677]: I1203 15:01:38.439987 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:01:38 crc kubenswrapper[4677]: I1203 15:01:38.440076 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" gracePeriod=600 Dec 03 15:01:39 crc kubenswrapper[4677]: I1203 15:01:39.003475 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" exitCode=0 Dec 03 15:01:39 crc kubenswrapper[4677]: I1203 15:01:39.004310 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d"} Dec 03 15:01:39 crc kubenswrapper[4677]: I1203 15:01:39.004422 4677 scope.go:117] "RemoveContainer" containerID="6aaa9b352a55266653b6997926c7a52e8b3205ac500a09ebc2a7d36ecaf352ee" Dec 03 15:01:39 crc kubenswrapper[4677]: E1203 15:01:39.346143 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:01:40 crc kubenswrapper[4677]: I1203 15:01:40.026994 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:01:40 crc kubenswrapper[4677]: E1203 15:01:40.027702 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:01:53 crc kubenswrapper[4677]: I1203 15:01:53.976327 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:01:53 crc kubenswrapper[4677]: E1203 15:01:53.977206 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:02:05 crc kubenswrapper[4677]: I1203 15:02:05.977081 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:02:05 crc kubenswrapper[4677]: E1203 15:02:05.978181 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:02:19 crc kubenswrapper[4677]: I1203 15:02:19.983192 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:02:19 crc kubenswrapper[4677]: E1203 15:02:19.984021 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:02:31 crc kubenswrapper[4677]: I1203 15:02:31.977049 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:02:31 crc kubenswrapper[4677]: E1203 15:02:31.979345 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:02:45 crc kubenswrapper[4677]: I1203 15:02:45.977292 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:02:45 crc kubenswrapper[4677]: E1203 15:02:45.978130 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:02:59 crc kubenswrapper[4677]: I1203 15:02:59.986886 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:02:59 crc kubenswrapper[4677]: E1203 15:02:59.987837 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:03:10 crc kubenswrapper[4677]: I1203 15:03:10.976536 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:03:10 crc kubenswrapper[4677]: E1203 15:03:10.977346 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:03:21 crc kubenswrapper[4677]: I1203 15:03:21.977923 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:03:21 crc kubenswrapper[4677]: E1203 15:03:21.978770 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:03:35 crc kubenswrapper[4677]: I1203 15:03:35.976879 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:03:35 crc kubenswrapper[4677]: E1203 15:03:35.978353 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:03:49 crc kubenswrapper[4677]: I1203 15:03:49.987991 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:03:49 crc kubenswrapper[4677]: E1203 15:03:49.990320 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:04:00 crc kubenswrapper[4677]: I1203 15:04:00.976459 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:04:00 crc kubenswrapper[4677]: E1203 15:04:00.977358 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:04:14 crc kubenswrapper[4677]: I1203 15:04:14.976437 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:04:14 crc kubenswrapper[4677]: E1203 15:04:14.977463 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:04:27 crc kubenswrapper[4677]: I1203 15:04:27.978748 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:04:27 crc kubenswrapper[4677]: E1203 15:04:27.980061 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:04:40 crc kubenswrapper[4677]: I1203 15:04:40.976744 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:04:40 crc kubenswrapper[4677]: E1203 15:04:40.977576 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:04:53 crc kubenswrapper[4677]: I1203 15:04:53.977487 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:04:53 crc kubenswrapper[4677]: E1203 15:04:53.978375 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:05:07 crc kubenswrapper[4677]: I1203 15:05:07.976748 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:05:07 crc kubenswrapper[4677]: E1203 15:05:07.977554 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:05:22 crc kubenswrapper[4677]: I1203 15:05:22.975764 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:05:22 crc kubenswrapper[4677]: E1203 15:05:22.976690 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:05:37 crc kubenswrapper[4677]: I1203 15:05:37.976735 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:05:37 crc kubenswrapper[4677]: E1203 15:05:37.977764 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:05:48 crc kubenswrapper[4677]: I1203 15:05:48.977019 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:05:48 crc kubenswrapper[4677]: E1203 15:05:48.978038 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:06:03 crc kubenswrapper[4677]: I1203 15:06:03.976478 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:06:03 crc kubenswrapper[4677]: E1203 15:06:03.977482 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:06:15 crc kubenswrapper[4677]: I1203 15:06:15.976195 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:06:15 crc kubenswrapper[4677]: E1203 15:06:15.977063 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:06:30 crc kubenswrapper[4677]: I1203 15:06:30.976478 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:06:30 crc kubenswrapper[4677]: E1203 15:06:30.977271 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:06:45 crc kubenswrapper[4677]: I1203 15:06:45.976700 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:06:46 crc kubenswrapper[4677]: I1203 15:06:46.342070 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"886fbe02f552cdfd3d8a20846a6e5149f750aafb759cf4a26f16cebc4d67067e"} Dec 03 15:07:16 crc kubenswrapper[4677]: I1203 15:07:16.741631 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-9cc645f55-v5xrn" podUID="5da71451-5801-4ce9-b177-cd847b635e24" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.078004 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m8622"] Dec 03 15:07:21 crc kubenswrapper[4677]: E1203 15:07:21.079175 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" containerName="keystone-cron" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.079196 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" containerName="keystone-cron" Dec 03 15:07:21 crc kubenswrapper[4677]: E1203 15:07:21.079212 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="registry-server" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.079220 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="registry-server" Dec 03 15:07:21 crc kubenswrapper[4677]: E1203 15:07:21.079234 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="extract-content" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.079241 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="extract-content" Dec 03 15:07:21 crc kubenswrapper[4677]: E1203 15:07:21.079295 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="extract-utilities" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.079303 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="extract-utilities" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.079589 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="b557bbbe-e8bc-493c-9b73-713665a1ca2b" containerName="registry-server" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.079621 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="da2e0a2c-7e17-4c0c-8cee-a2e2cee92632" containerName="keystone-cron" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.081417 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.096652 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m8622"] Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.205477 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8bpc\" (UniqueName: \"kubernetes.io/projected/9ff4bc81-de65-4061-82b1-6c21016159f6-kube-api-access-z8bpc\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.205696 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-catalog-content\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.205800 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-utilities\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.307788 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-catalog-content\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.307911 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-utilities\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.307943 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8bpc\" (UniqueName: \"kubernetes.io/projected/9ff4bc81-de65-4061-82b1-6c21016159f6-kube-api-access-z8bpc\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.308468 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-catalog-content\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.308502 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-utilities\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.333909 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8bpc\" (UniqueName: \"kubernetes.io/projected/9ff4bc81-de65-4061-82b1-6c21016159f6-kube-api-access-z8bpc\") pod \"redhat-marketplace-m8622\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.422914 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:21 crc kubenswrapper[4677]: I1203 15:07:21.919979 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m8622"] Dec 03 15:07:22 crc kubenswrapper[4677]: I1203 15:07:22.792126 4677 generic.go:334] "Generic (PLEG): container finished" podID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerID="001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b" exitCode=0 Dec 03 15:07:22 crc kubenswrapper[4677]: I1203 15:07:22.792215 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8622" event={"ID":"9ff4bc81-de65-4061-82b1-6c21016159f6","Type":"ContainerDied","Data":"001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b"} Dec 03 15:07:22 crc kubenswrapper[4677]: I1203 15:07:22.792601 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8622" event={"ID":"9ff4bc81-de65-4061-82b1-6c21016159f6","Type":"ContainerStarted","Data":"99e2b424c08a6b99915969dbec250ab9a6f93f6dea6ac7df05454797a34af407"} Dec 03 15:07:22 crc kubenswrapper[4677]: I1203 15:07:22.795185 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:07:24 crc kubenswrapper[4677]: I1203 15:07:24.843732 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8622" event={"ID":"9ff4bc81-de65-4061-82b1-6c21016159f6","Type":"ContainerStarted","Data":"48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73"} Dec 03 15:07:25 crc kubenswrapper[4677]: I1203 15:07:25.854176 4677 generic.go:334] "Generic (PLEG): container finished" podID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerID="48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73" exitCode=0 Dec 03 15:07:25 crc kubenswrapper[4677]: I1203 15:07:25.854298 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8622" event={"ID":"9ff4bc81-de65-4061-82b1-6c21016159f6","Type":"ContainerDied","Data":"48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73"} Dec 03 15:07:26 crc kubenswrapper[4677]: I1203 15:07:26.872663 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8622" event={"ID":"9ff4bc81-de65-4061-82b1-6c21016159f6","Type":"ContainerStarted","Data":"396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678"} Dec 03 15:07:26 crc kubenswrapper[4677]: I1203 15:07:26.899958 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m8622" podStartSLOduration=2.293137859 podStartE2EDuration="5.899921492s" podCreationTimestamp="2025-12-03 15:07:21 +0000 UTC" firstStartedPulling="2025-12-03 15:07:22.79452568 +0000 UTC m=+4833.540858175" lastFinishedPulling="2025-12-03 15:07:26.401309353 +0000 UTC m=+4837.147641808" observedRunningTime="2025-12-03 15:07:26.895799188 +0000 UTC m=+4837.642131653" watchObservedRunningTime="2025-12-03 15:07:26.899921492 +0000 UTC m=+4837.646253947" Dec 03 15:07:31 crc kubenswrapper[4677]: I1203 15:07:31.423082 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:31 crc kubenswrapper[4677]: I1203 15:07:31.423671 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:31 crc kubenswrapper[4677]: I1203 15:07:31.476857 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:32 crc kubenswrapper[4677]: I1203 15:07:32.010458 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:32 crc kubenswrapper[4677]: I1203 15:07:32.067756 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m8622"] Dec 03 15:07:33 crc kubenswrapper[4677]: I1203 15:07:33.951006 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m8622" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="registry-server" containerID="cri-o://396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678" gracePeriod=2 Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.466194 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.518180 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-utilities\") pod \"9ff4bc81-de65-4061-82b1-6c21016159f6\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.518336 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8bpc\" (UniqueName: \"kubernetes.io/projected/9ff4bc81-de65-4061-82b1-6c21016159f6-kube-api-access-z8bpc\") pod \"9ff4bc81-de65-4061-82b1-6c21016159f6\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.518430 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-catalog-content\") pod \"9ff4bc81-de65-4061-82b1-6c21016159f6\" (UID: \"9ff4bc81-de65-4061-82b1-6c21016159f6\") " Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.519303 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-utilities" (OuterVolumeSpecName: "utilities") pod "9ff4bc81-de65-4061-82b1-6c21016159f6" (UID: "9ff4bc81-de65-4061-82b1-6c21016159f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.528290 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff4bc81-de65-4061-82b1-6c21016159f6-kube-api-access-z8bpc" (OuterVolumeSpecName: "kube-api-access-z8bpc") pod "9ff4bc81-de65-4061-82b1-6c21016159f6" (UID: "9ff4bc81-de65-4061-82b1-6c21016159f6"). InnerVolumeSpecName "kube-api-access-z8bpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.549243 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ff4bc81-de65-4061-82b1-6c21016159f6" (UID: "9ff4bc81-de65-4061-82b1-6c21016159f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.621588 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8bpc\" (UniqueName: \"kubernetes.io/projected/9ff4bc81-de65-4061-82b1-6c21016159f6-kube-api-access-z8bpc\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.621634 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.621647 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff4bc81-de65-4061-82b1-6c21016159f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.964833 4677 generic.go:334] "Generic (PLEG): container finished" podID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerID="396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678" exitCode=0 Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.964897 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m8622" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.964930 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8622" event={"ID":"9ff4bc81-de65-4061-82b1-6c21016159f6","Type":"ContainerDied","Data":"396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678"} Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.965274 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m8622" event={"ID":"9ff4bc81-de65-4061-82b1-6c21016159f6","Type":"ContainerDied","Data":"99e2b424c08a6b99915969dbec250ab9a6f93f6dea6ac7df05454797a34af407"} Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.965316 4677 scope.go:117] "RemoveContainer" containerID="396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678" Dec 03 15:07:34 crc kubenswrapper[4677]: I1203 15:07:34.989854 4677 scope.go:117] "RemoveContainer" containerID="48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.013032 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m8622"] Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.024941 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m8622"] Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.038571 4677 scope.go:117] "RemoveContainer" containerID="001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.097598 4677 scope.go:117] "RemoveContainer" containerID="396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678" Dec 03 15:07:35 crc kubenswrapper[4677]: E1203 15:07:35.099172 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678\": container with ID starting with 396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678 not found: ID does not exist" containerID="396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.099256 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678"} err="failed to get container status \"396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678\": rpc error: code = NotFound desc = could not find container \"396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678\": container with ID starting with 396e7a3a4d01335af8a93f27cff02aca44e4a6db68ec5719fddd84d5e3b0d678 not found: ID does not exist" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.099291 4677 scope.go:117] "RemoveContainer" containerID="48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73" Dec 03 15:07:35 crc kubenswrapper[4677]: E1203 15:07:35.099811 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73\": container with ID starting with 48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73 not found: ID does not exist" containerID="48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.099918 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73"} err="failed to get container status \"48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73\": rpc error: code = NotFound desc = could not find container \"48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73\": container with ID starting with 48dbd608977632dd9c4ca02c280cfa393fc716dcccef976acc1fc0488daa1e73 not found: ID does not exist" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.100019 4677 scope.go:117] "RemoveContainer" containerID="001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b" Dec 03 15:07:35 crc kubenswrapper[4677]: E1203 15:07:35.100938 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b\": container with ID starting with 001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b not found: ID does not exist" containerID="001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.100984 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b"} err="failed to get container status \"001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b\": rpc error: code = NotFound desc = could not find container \"001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b\": container with ID starting with 001cb51a80738cade7628971f33eb637168109cd2dae5227ff6c7b88c1cd872b not found: ID does not exist" Dec 03 15:07:35 crc kubenswrapper[4677]: I1203 15:07:35.988887 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" path="/var/lib/kubelet/pods/9ff4bc81-de65-4061-82b1-6c21016159f6/volumes" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.809474 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-frz49"] Dec 03 15:07:47 crc kubenswrapper[4677]: E1203 15:07:47.810628 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="registry-server" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.810647 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="registry-server" Dec 03 15:07:47 crc kubenswrapper[4677]: E1203 15:07:47.810690 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="extract-content" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.810699 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="extract-content" Dec 03 15:07:47 crc kubenswrapper[4677]: E1203 15:07:47.810722 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="extract-utilities" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.810731 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="extract-utilities" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.810992 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ff4bc81-de65-4061-82b1-6c21016159f6" containerName="registry-server" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.812877 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.818648 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-frz49"] Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.939771 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-catalog-content\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.939987 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dz2n\" (UniqueName: \"kubernetes.io/projected/afb9c3f4-b3fd-43b6-b044-61e52e2120de-kube-api-access-7dz2n\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:47 crc kubenswrapper[4677]: I1203 15:07:47.940325 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-utilities\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:48 crc kubenswrapper[4677]: I1203 15:07:48.042728 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-utilities\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:48 crc kubenswrapper[4677]: I1203 15:07:48.042838 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-catalog-content\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:48 crc kubenswrapper[4677]: I1203 15:07:48.042900 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dz2n\" (UniqueName: \"kubernetes.io/projected/afb9c3f4-b3fd-43b6-b044-61e52e2120de-kube-api-access-7dz2n\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:48 crc kubenswrapper[4677]: I1203 15:07:48.043546 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-catalog-content\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:48 crc kubenswrapper[4677]: I1203 15:07:48.043577 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-utilities\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:48 crc kubenswrapper[4677]: I1203 15:07:48.602640 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dz2n\" (UniqueName: \"kubernetes.io/projected/afb9c3f4-b3fd-43b6-b044-61e52e2120de-kube-api-access-7dz2n\") pod \"certified-operators-frz49\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:48 crc kubenswrapper[4677]: I1203 15:07:48.740271 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:49 crc kubenswrapper[4677]: I1203 15:07:49.218619 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-frz49"] Dec 03 15:07:50 crc kubenswrapper[4677]: I1203 15:07:50.127100 4677 generic.go:334] "Generic (PLEG): container finished" podID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerID="e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854" exitCode=0 Dec 03 15:07:50 crc kubenswrapper[4677]: I1203 15:07:50.127267 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frz49" event={"ID":"afb9c3f4-b3fd-43b6-b044-61e52e2120de","Type":"ContainerDied","Data":"e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854"} Dec 03 15:07:50 crc kubenswrapper[4677]: I1203 15:07:50.127378 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frz49" event={"ID":"afb9c3f4-b3fd-43b6-b044-61e52e2120de","Type":"ContainerStarted","Data":"d45de018ad420183362697202016f9e8e22f0867a29e17e201d15d8d5f26b16d"} Dec 03 15:07:52 crc kubenswrapper[4677]: I1203 15:07:52.168584 4677 generic.go:334] "Generic (PLEG): container finished" podID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerID="de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70" exitCode=0 Dec 03 15:07:52 crc kubenswrapper[4677]: I1203 15:07:52.168709 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frz49" event={"ID":"afb9c3f4-b3fd-43b6-b044-61e52e2120de","Type":"ContainerDied","Data":"de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70"} Dec 03 15:07:53 crc kubenswrapper[4677]: I1203 15:07:53.184288 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frz49" event={"ID":"afb9c3f4-b3fd-43b6-b044-61e52e2120de","Type":"ContainerStarted","Data":"222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a"} Dec 03 15:07:53 crc kubenswrapper[4677]: I1203 15:07:53.206041 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-frz49" podStartSLOduration=3.690045282 podStartE2EDuration="6.206022015s" podCreationTimestamp="2025-12-03 15:07:47 +0000 UTC" firstStartedPulling="2025-12-03 15:07:50.129348554 +0000 UTC m=+4860.875681009" lastFinishedPulling="2025-12-03 15:07:52.645325247 +0000 UTC m=+4863.391657742" observedRunningTime="2025-12-03 15:07:53.200650358 +0000 UTC m=+4863.946982833" watchObservedRunningTime="2025-12-03 15:07:53.206022015 +0000 UTC m=+4863.952354460" Dec 03 15:07:58 crc kubenswrapper[4677]: I1203 15:07:58.740899 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:58 crc kubenswrapper[4677]: I1203 15:07:58.741809 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:58 crc kubenswrapper[4677]: I1203 15:07:58.810286 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:59 crc kubenswrapper[4677]: I1203 15:07:59.334613 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:07:59 crc kubenswrapper[4677]: I1203 15:07:59.402079 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-frz49"] Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.292832 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-frz49" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="registry-server" containerID="cri-o://222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a" gracePeriod=2 Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.805690 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.860647 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-catalog-content\") pod \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.860734 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dz2n\" (UniqueName: \"kubernetes.io/projected/afb9c3f4-b3fd-43b6-b044-61e52e2120de-kube-api-access-7dz2n\") pod \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.860807 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-utilities\") pod \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\" (UID: \"afb9c3f4-b3fd-43b6-b044-61e52e2120de\") " Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.861927 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-utilities" (OuterVolumeSpecName: "utilities") pod "afb9c3f4-b3fd-43b6-b044-61e52e2120de" (UID: "afb9c3f4-b3fd-43b6-b044-61e52e2120de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.868273 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afb9c3f4-b3fd-43b6-b044-61e52e2120de-kube-api-access-7dz2n" (OuterVolumeSpecName: "kube-api-access-7dz2n") pod "afb9c3f4-b3fd-43b6-b044-61e52e2120de" (UID: "afb9c3f4-b3fd-43b6-b044-61e52e2120de"). InnerVolumeSpecName "kube-api-access-7dz2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.915162 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afb9c3f4-b3fd-43b6-b044-61e52e2120de" (UID: "afb9c3f4-b3fd-43b6-b044-61e52e2120de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.963645 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.963693 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb9c3f4-b3fd-43b6-b044-61e52e2120de-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:08:01 crc kubenswrapper[4677]: I1203 15:08:01.963709 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dz2n\" (UniqueName: \"kubernetes.io/projected/afb9c3f4-b3fd-43b6-b044-61e52e2120de-kube-api-access-7dz2n\") on node \"crc\" DevicePath \"\"" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.308977 4677 generic.go:334] "Generic (PLEG): container finished" podID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerID="222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a" exitCode=0 Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.309038 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frz49" event={"ID":"afb9c3f4-b3fd-43b6-b044-61e52e2120de","Type":"ContainerDied","Data":"222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a"} Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.309084 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frz49" event={"ID":"afb9c3f4-b3fd-43b6-b044-61e52e2120de","Type":"ContainerDied","Data":"d45de018ad420183362697202016f9e8e22f0867a29e17e201d15d8d5f26b16d"} Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.309083 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frz49" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.309108 4677 scope.go:117] "RemoveContainer" containerID="222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.349047 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-frz49"] Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.350239 4677 scope.go:117] "RemoveContainer" containerID="de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.364097 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-frz49"] Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.391595 4677 scope.go:117] "RemoveContainer" containerID="e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.451582 4677 scope.go:117] "RemoveContainer" containerID="222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a" Dec 03 15:08:02 crc kubenswrapper[4677]: E1203 15:08:02.452315 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a\": container with ID starting with 222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a not found: ID does not exist" containerID="222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.452407 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a"} err="failed to get container status \"222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a\": rpc error: code = NotFound desc = could not find container \"222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a\": container with ID starting with 222409a4ad794d28226d98ad9872175f5c22948172af73ba11c6eb194350bc9a not found: ID does not exist" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.452471 4677 scope.go:117] "RemoveContainer" containerID="de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70" Dec 03 15:08:02 crc kubenswrapper[4677]: E1203 15:08:02.453332 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70\": container with ID starting with de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70 not found: ID does not exist" containerID="de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.453404 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70"} err="failed to get container status \"de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70\": rpc error: code = NotFound desc = could not find container \"de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70\": container with ID starting with de4f18d3e3af7a022c2006b51a182822bd54c461ffb71b20573ced222fcfde70 not found: ID does not exist" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.453458 4677 scope.go:117] "RemoveContainer" containerID="e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854" Dec 03 15:08:02 crc kubenswrapper[4677]: E1203 15:08:02.454340 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854\": container with ID starting with e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854 not found: ID does not exist" containerID="e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854" Dec 03 15:08:02 crc kubenswrapper[4677]: I1203 15:08:02.454722 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854"} err="failed to get container status \"e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854\": rpc error: code = NotFound desc = could not find container \"e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854\": container with ID starting with e35dce1fb031fe98b1009b5d4def587620868f126a7840cec59531556176e854 not found: ID does not exist" Dec 03 15:08:03 crc kubenswrapper[4677]: I1203 15:08:03.991739 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" path="/var/lib/kubelet/pods/afb9c3f4-b3fd-43b6-b044-61e52e2120de/volumes" Dec 03 15:09:08 crc kubenswrapper[4677]: I1203 15:09:08.437374 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:09:08 crc kubenswrapper[4677]: I1203 15:09:08.438018 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:09:38 crc kubenswrapper[4677]: I1203 15:09:38.437605 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:09:38 crc kubenswrapper[4677]: I1203 15:09:38.438261 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:09:55 crc kubenswrapper[4677]: I1203 15:09:55.823144 4677 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5dcf4b8dc4-2f58c" podUID="9427b27d-5375-4914-9481-fbbd6c1f9e21" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.57:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.437513 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.438225 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.438290 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.439345 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"886fbe02f552cdfd3d8a20846a6e5149f750aafb759cf4a26f16cebc4d67067e"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.439420 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://886fbe02f552cdfd3d8a20846a6e5149f750aafb759cf4a26f16cebc4d67067e" gracePeriod=600 Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.729212 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="886fbe02f552cdfd3d8a20846a6e5149f750aafb759cf4a26f16cebc4d67067e" exitCode=0 Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.729304 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"886fbe02f552cdfd3d8a20846a6e5149f750aafb759cf4a26f16cebc4d67067e"} Dec 03 15:10:08 crc kubenswrapper[4677]: I1203 15:10:08.730304 4677 scope.go:117] "RemoveContainer" containerID="f3a40fe2d50e8bf5ef4c32aa2a6f130ceca324d11b7a3142e919f3b8a2277d0d" Dec 03 15:10:09 crc kubenswrapper[4677]: I1203 15:10:09.746846 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce"} Dec 03 15:12:08 crc kubenswrapper[4677]: I1203 15:12:08.437989 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:12:08 crc kubenswrapper[4677]: I1203 15:12:08.438710 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:12:38 crc kubenswrapper[4677]: I1203 15:12:38.437938 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:12:38 crc kubenswrapper[4677]: I1203 15:12:38.438531 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.436839 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.437341 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.437414 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.438123 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.438177 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" gracePeriod=600 Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.783320 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" exitCode=0 Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.783405 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce"} Dec 03 15:13:08 crc kubenswrapper[4677]: I1203 15:13:08.783568 4677 scope.go:117] "RemoveContainer" containerID="886fbe02f552cdfd3d8a20846a6e5149f750aafb759cf4a26f16cebc4d67067e" Dec 03 15:13:09 crc kubenswrapper[4677]: E1203 15:13:09.076985 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:13:09 crc kubenswrapper[4677]: I1203 15:13:09.800008 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:13:09 crc kubenswrapper[4677]: E1203 15:13:09.800656 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:13:19 crc kubenswrapper[4677]: E1203 15:13:19.034819 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 03 15:13:22 crc kubenswrapper[4677]: I1203 15:13:22.977294 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:13:22 crc kubenswrapper[4677]: E1203 15:13:22.978509 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:13:33 crc kubenswrapper[4677]: I1203 15:13:33.977294 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:13:33 crc kubenswrapper[4677]: E1203 15:13:33.978771 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:13:44 crc kubenswrapper[4677]: I1203 15:13:44.977432 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:13:44 crc kubenswrapper[4677]: E1203 15:13:44.978656 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:13:58 crc kubenswrapper[4677]: I1203 15:13:58.976583 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:13:58 crc kubenswrapper[4677]: E1203 15:13:58.979095 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:14:09 crc kubenswrapper[4677]: I1203 15:14:09.989122 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:14:09 crc kubenswrapper[4677]: E1203 15:14:09.990173 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:14:20 crc kubenswrapper[4677]: I1203 15:14:20.976857 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:14:20 crc kubenswrapper[4677]: E1203 15:14:20.977729 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:14:32 crc kubenswrapper[4677]: I1203 15:14:32.975994 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:14:32 crc kubenswrapper[4677]: E1203 15:14:32.976823 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.868014 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cddzp"] Dec 03 15:14:38 crc kubenswrapper[4677]: E1203 15:14:38.869175 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="extract-content" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.869196 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="extract-content" Dec 03 15:14:38 crc kubenswrapper[4677]: E1203 15:14:38.869231 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="extract-utilities" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.869238 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="extract-utilities" Dec 03 15:14:38 crc kubenswrapper[4677]: E1203 15:14:38.869248 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="registry-server" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.869254 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="registry-server" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.869510 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="afb9c3f4-b3fd-43b6-b044-61e52e2120de" containerName="registry-server" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.870940 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.893601 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cddzp"] Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.948825 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5gx8\" (UniqueName: \"kubernetes.io/projected/21c1c4d5-0c44-4edc-9271-445194132d6e-kube-api-access-x5gx8\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.948894 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-catalog-content\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:38 crc kubenswrapper[4677]: I1203 15:14:38.949035 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-utilities\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.050894 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-utilities\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.051083 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5gx8\" (UniqueName: \"kubernetes.io/projected/21c1c4d5-0c44-4edc-9271-445194132d6e-kube-api-access-x5gx8\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.051115 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-catalog-content\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.052008 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-utilities\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.052171 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-catalog-content\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.062416 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hmhnx"] Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.064504 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.080882 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5gx8\" (UniqueName: \"kubernetes.io/projected/21c1c4d5-0c44-4edc-9271-445194132d6e-kube-api-access-x5gx8\") pod \"redhat-operators-cddzp\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.082395 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hmhnx"] Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.152406 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l7nh\" (UniqueName: \"kubernetes.io/projected/2977934a-8718-4a2e-b22d-fc71ae826f14-kube-api-access-8l7nh\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.152557 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-utilities\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.152640 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-catalog-content\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.189724 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.255543 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-utilities\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.256388 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-utilities\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.257250 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-catalog-content\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.257343 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l7nh\" (UniqueName: \"kubernetes.io/projected/2977934a-8718-4a2e-b22d-fc71ae826f14-kube-api-access-8l7nh\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.257825 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-catalog-content\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.288917 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l7nh\" (UniqueName: \"kubernetes.io/projected/2977934a-8718-4a2e-b22d-fc71ae826f14-kube-api-access-8l7nh\") pod \"community-operators-hmhnx\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.445547 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:39 crc kubenswrapper[4677]: I1203 15:14:39.842298 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cddzp"] Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.161778 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hmhnx"] Dec 03 15:14:40 crc kubenswrapper[4677]: W1203 15:14:40.162871 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2977934a_8718_4a2e_b22d_fc71ae826f14.slice/crio-afdaac3dd7a201a2813ab68efeae9d83fd560498bf577ea6461fb7b16c1033ab WatchSource:0}: Error finding container afdaac3dd7a201a2813ab68efeae9d83fd560498bf577ea6461fb7b16c1033ab: Status 404 returned error can't find the container with id afdaac3dd7a201a2813ab68efeae9d83fd560498bf577ea6461fb7b16c1033ab Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.835013 4677 generic.go:334] "Generic (PLEG): container finished" podID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerID="d9992bd0ddc46128e5f8913fde47ea86aee44e3e5ff85295d598755cbf7ab60e" exitCode=0 Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.835117 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hmhnx" event={"ID":"2977934a-8718-4a2e-b22d-fc71ae826f14","Type":"ContainerDied","Data":"d9992bd0ddc46128e5f8913fde47ea86aee44e3e5ff85295d598755cbf7ab60e"} Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.835459 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hmhnx" event={"ID":"2977934a-8718-4a2e-b22d-fc71ae826f14","Type":"ContainerStarted","Data":"afdaac3dd7a201a2813ab68efeae9d83fd560498bf577ea6461fb7b16c1033ab"} Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.836733 4677 generic.go:334] "Generic (PLEG): container finished" podID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerID="247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326" exitCode=0 Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.836783 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cddzp" event={"ID":"21c1c4d5-0c44-4edc-9271-445194132d6e","Type":"ContainerDied","Data":"247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326"} Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.836812 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cddzp" event={"ID":"21c1c4d5-0c44-4edc-9271-445194132d6e","Type":"ContainerStarted","Data":"64bb9bd515baba8e117f07e63052dfe0fe12244c01d8512c7b594e9210223f56"} Dec 03 15:14:40 crc kubenswrapper[4677]: I1203 15:14:40.837526 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:14:41 crc kubenswrapper[4677]: I1203 15:14:41.851502 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cddzp" event={"ID":"21c1c4d5-0c44-4edc-9271-445194132d6e","Type":"ContainerStarted","Data":"4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc"} Dec 03 15:14:41 crc kubenswrapper[4677]: I1203 15:14:41.858831 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hmhnx" event={"ID":"2977934a-8718-4a2e-b22d-fc71ae826f14","Type":"ContainerStarted","Data":"6b4178e265459b57e5a703705af3e1c20b756e6731475bd1b2ee44a58005b477"} Dec 03 15:14:44 crc kubenswrapper[4677]: I1203 15:14:44.896076 4677 generic.go:334] "Generic (PLEG): container finished" podID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerID="6b4178e265459b57e5a703705af3e1c20b756e6731475bd1b2ee44a58005b477" exitCode=0 Dec 03 15:14:44 crc kubenswrapper[4677]: I1203 15:14:44.896178 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hmhnx" event={"ID":"2977934a-8718-4a2e-b22d-fc71ae826f14","Type":"ContainerDied","Data":"6b4178e265459b57e5a703705af3e1c20b756e6731475bd1b2ee44a58005b477"} Dec 03 15:14:46 crc kubenswrapper[4677]: I1203 15:14:46.922834 4677 generic.go:334] "Generic (PLEG): container finished" podID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerID="4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc" exitCode=0 Dec 03 15:14:46 crc kubenswrapper[4677]: I1203 15:14:46.922913 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cddzp" event={"ID":"21c1c4d5-0c44-4edc-9271-445194132d6e","Type":"ContainerDied","Data":"4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc"} Dec 03 15:14:46 crc kubenswrapper[4677]: I1203 15:14:46.926873 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hmhnx" event={"ID":"2977934a-8718-4a2e-b22d-fc71ae826f14","Type":"ContainerStarted","Data":"93f09dc5ec5020d34261c7ad1510ff12829b93ac17374242061d05223c03b9de"} Dec 03 15:14:46 crc kubenswrapper[4677]: I1203 15:14:46.975823 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hmhnx" podStartSLOduration=2.547494959 podStartE2EDuration="7.975801893s" podCreationTimestamp="2025-12-03 15:14:39 +0000 UTC" firstStartedPulling="2025-12-03 15:14:40.837310618 +0000 UTC m=+5271.583643073" lastFinishedPulling="2025-12-03 15:14:46.265617512 +0000 UTC m=+5277.011950007" observedRunningTime="2025-12-03 15:14:46.970091627 +0000 UTC m=+5277.716424102" watchObservedRunningTime="2025-12-03 15:14:46.975801893 +0000 UTC m=+5277.722134348" Dec 03 15:14:46 crc kubenswrapper[4677]: I1203 15:14:46.977118 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:14:46 crc kubenswrapper[4677]: E1203 15:14:46.977607 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:14:48 crc kubenswrapper[4677]: I1203 15:14:48.954767 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cddzp" event={"ID":"21c1c4d5-0c44-4edc-9271-445194132d6e","Type":"ContainerStarted","Data":"954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763"} Dec 03 15:14:48 crc kubenswrapper[4677]: I1203 15:14:48.986771 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cddzp" podStartSLOduration=4.078035511 podStartE2EDuration="10.986740051s" podCreationTimestamp="2025-12-03 15:14:38 +0000 UTC" firstStartedPulling="2025-12-03 15:14:40.839298281 +0000 UTC m=+5271.585630736" lastFinishedPulling="2025-12-03 15:14:47.748002821 +0000 UTC m=+5278.494335276" observedRunningTime="2025-12-03 15:14:48.976079899 +0000 UTC m=+5279.722412534" watchObservedRunningTime="2025-12-03 15:14:48.986740051 +0000 UTC m=+5279.733072506" Dec 03 15:14:49 crc kubenswrapper[4677]: I1203 15:14:49.190374 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:49 crc kubenswrapper[4677]: I1203 15:14:49.190556 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:49 crc kubenswrapper[4677]: I1203 15:14:49.446521 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:49 crc kubenswrapper[4677]: I1203 15:14:49.446583 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:49 crc kubenswrapper[4677]: I1203 15:14:49.506359 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:14:50 crc kubenswrapper[4677]: I1203 15:14:50.267703 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cddzp" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="registry-server" probeResult="failure" output=< Dec 03 15:14:50 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 15:14:50 crc kubenswrapper[4677]: > Dec 03 15:14:58 crc kubenswrapper[4677]: I1203 15:14:58.976228 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:14:58 crc kubenswrapper[4677]: E1203 15:14:58.977216 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:14:59 crc kubenswrapper[4677]: I1203 15:14:59.250199 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:59 crc kubenswrapper[4677]: I1203 15:14:59.311482 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:14:59 crc kubenswrapper[4677]: I1203 15:14:59.498531 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cddzp"] Dec 03 15:14:59 crc kubenswrapper[4677]: I1203 15:14:59.507637 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.163073 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q"] Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.164836 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.171204 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.171508 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.172723 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q"] Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.307494 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhfpr\" (UniqueName: \"kubernetes.io/projected/4aecb6a0-009d-43bb-a67f-61e0c2978610-kube-api-access-rhfpr\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.307587 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aecb6a0-009d-43bb-a67f-61e0c2978610-secret-volume\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.307715 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aecb6a0-009d-43bb-a67f-61e0c2978610-config-volume\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.409592 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhfpr\" (UniqueName: \"kubernetes.io/projected/4aecb6a0-009d-43bb-a67f-61e0c2978610-kube-api-access-rhfpr\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.409669 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aecb6a0-009d-43bb-a67f-61e0c2978610-secret-volume\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.409762 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aecb6a0-009d-43bb-a67f-61e0c2978610-config-volume\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.411002 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aecb6a0-009d-43bb-a67f-61e0c2978610-config-volume\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.430435 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aecb6a0-009d-43bb-a67f-61e0c2978610-secret-volume\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.438062 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhfpr\" (UniqueName: \"kubernetes.io/projected/4aecb6a0-009d-43bb-a67f-61e0c2978610-kube-api-access-rhfpr\") pod \"collect-profiles-29412915-nm44q\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.492158 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:00 crc kubenswrapper[4677]: I1203 15:15:00.980756 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q"] Dec 03 15:15:00 crc kubenswrapper[4677]: W1203 15:15:00.987141 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aecb6a0_009d_43bb_a67f_61e0c2978610.slice/crio-650b81cbc074cb9ceaa7af6c602c9de206f302fde96fd2daf6650cb925799ef9 WatchSource:0}: Error finding container 650b81cbc074cb9ceaa7af6c602c9de206f302fde96fd2daf6650cb925799ef9: Status 404 returned error can't find the container with id 650b81cbc074cb9ceaa7af6c602c9de206f302fde96fd2daf6650cb925799ef9 Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.089074 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" event={"ID":"4aecb6a0-009d-43bb-a67f-61e0c2978610","Type":"ContainerStarted","Data":"650b81cbc074cb9ceaa7af6c602c9de206f302fde96fd2daf6650cb925799ef9"} Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.089380 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cddzp" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="registry-server" containerID="cri-o://954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763" gracePeriod=2 Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.763166 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.839283 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5gx8\" (UniqueName: \"kubernetes.io/projected/21c1c4d5-0c44-4edc-9271-445194132d6e-kube-api-access-x5gx8\") pod \"21c1c4d5-0c44-4edc-9271-445194132d6e\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.839440 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-catalog-content\") pod \"21c1c4d5-0c44-4edc-9271-445194132d6e\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.839650 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-utilities\") pod \"21c1c4d5-0c44-4edc-9271-445194132d6e\" (UID: \"21c1c4d5-0c44-4edc-9271-445194132d6e\") " Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.840477 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-utilities" (OuterVolumeSpecName: "utilities") pod "21c1c4d5-0c44-4edc-9271-445194132d6e" (UID: "21c1c4d5-0c44-4edc-9271-445194132d6e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.846932 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21c1c4d5-0c44-4edc-9271-445194132d6e-kube-api-access-x5gx8" (OuterVolumeSpecName: "kube-api-access-x5gx8") pod "21c1c4d5-0c44-4edc-9271-445194132d6e" (UID: "21c1c4d5-0c44-4edc-9271-445194132d6e"). InnerVolumeSpecName "kube-api-access-x5gx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.895731 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hmhnx"] Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.896036 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hmhnx" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="registry-server" containerID="cri-o://93f09dc5ec5020d34261c7ad1510ff12829b93ac17374242061d05223c03b9de" gracePeriod=2 Dec 03 15:15:01 crc kubenswrapper[4677]: E1203 15:15:01.917192 4677 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aecb6a0_009d_43bb_a67f_61e0c2978610.slice/crio-conmon-7bf6e91f873ecb86aa8539ba51557e2b82ba375c1ed8c7132d63630dbe94e1fa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aecb6a0_009d_43bb_a67f_61e0c2978610.slice/crio-7bf6e91f873ecb86aa8539ba51557e2b82ba375c1ed8c7132d63630dbe94e1fa.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.942905 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5gx8\" (UniqueName: \"kubernetes.io/projected/21c1c4d5-0c44-4edc-9271-445194132d6e-kube-api-access-x5gx8\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.942969 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:01 crc kubenswrapper[4677]: I1203 15:15:01.988785 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21c1c4d5-0c44-4edc-9271-445194132d6e" (UID: "21c1c4d5-0c44-4edc-9271-445194132d6e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.045108 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c1c4d5-0c44-4edc-9271-445194132d6e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.113767 4677 generic.go:334] "Generic (PLEG): container finished" podID="4aecb6a0-009d-43bb-a67f-61e0c2978610" containerID="7bf6e91f873ecb86aa8539ba51557e2b82ba375c1ed8c7132d63630dbe94e1fa" exitCode=0 Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.113844 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" event={"ID":"4aecb6a0-009d-43bb-a67f-61e0c2978610","Type":"ContainerDied","Data":"7bf6e91f873ecb86aa8539ba51557e2b82ba375c1ed8c7132d63630dbe94e1fa"} Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.118567 4677 generic.go:334] "Generic (PLEG): container finished" podID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerID="954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763" exitCode=0 Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.118666 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cddzp" event={"ID":"21c1c4d5-0c44-4edc-9271-445194132d6e","Type":"ContainerDied","Data":"954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763"} Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.118707 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cddzp" event={"ID":"21c1c4d5-0c44-4edc-9271-445194132d6e","Type":"ContainerDied","Data":"64bb9bd515baba8e117f07e63052dfe0fe12244c01d8512c7b594e9210223f56"} Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.118735 4677 scope.go:117] "RemoveContainer" containerID="954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.119026 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cddzp" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.129627 4677 generic.go:334] "Generic (PLEG): container finished" podID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerID="93f09dc5ec5020d34261c7ad1510ff12829b93ac17374242061d05223c03b9de" exitCode=0 Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.129688 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hmhnx" event={"ID":"2977934a-8718-4a2e-b22d-fc71ae826f14","Type":"ContainerDied","Data":"93f09dc5ec5020d34261c7ad1510ff12829b93ac17374242061d05223c03b9de"} Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.178601 4677 scope.go:117] "RemoveContainer" containerID="4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.187162 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cddzp"] Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.196355 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cddzp"] Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.221854 4677 scope.go:117] "RemoveContainer" containerID="247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.240917 4677 scope.go:117] "RemoveContainer" containerID="954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763" Dec 03 15:15:02 crc kubenswrapper[4677]: E1203 15:15:02.241303 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763\": container with ID starting with 954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763 not found: ID does not exist" containerID="954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.241366 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763"} err="failed to get container status \"954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763\": rpc error: code = NotFound desc = could not find container \"954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763\": container with ID starting with 954b2aba5d52f5253e590d5cdf7c1d2abb42ee9e61d545a5f7ae5bdd000f5763 not found: ID does not exist" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.241412 4677 scope.go:117] "RemoveContainer" containerID="4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc" Dec 03 15:15:02 crc kubenswrapper[4677]: E1203 15:15:02.241937 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc\": container with ID starting with 4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc not found: ID does not exist" containerID="4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.242031 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc"} err="failed to get container status \"4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc\": rpc error: code = NotFound desc = could not find container \"4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc\": container with ID starting with 4f5403b263584333ca4b8f6c8dfb3a396a04d82c9ea16c89f1f8d0d956b238cc not found: ID does not exist" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.242097 4677 scope.go:117] "RemoveContainer" containerID="247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326" Dec 03 15:15:02 crc kubenswrapper[4677]: E1203 15:15:02.242468 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326\": container with ID starting with 247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326 not found: ID does not exist" containerID="247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.242507 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326"} err="failed to get container status \"247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326\": rpc error: code = NotFound desc = could not find container \"247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326\": container with ID starting with 247d5c32b86f2820c1c6e8724b417a358dfb02e50d2cc24caee9980eeb4ad326 not found: ID does not exist" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.270739 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.355854 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-catalog-content\") pod \"2977934a-8718-4a2e-b22d-fc71ae826f14\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.355987 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-utilities\") pod \"2977934a-8718-4a2e-b22d-fc71ae826f14\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.356066 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l7nh\" (UniqueName: \"kubernetes.io/projected/2977934a-8718-4a2e-b22d-fc71ae826f14-kube-api-access-8l7nh\") pod \"2977934a-8718-4a2e-b22d-fc71ae826f14\" (UID: \"2977934a-8718-4a2e-b22d-fc71ae826f14\") " Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.356592 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-utilities" (OuterVolumeSpecName: "utilities") pod "2977934a-8718-4a2e-b22d-fc71ae826f14" (UID: "2977934a-8718-4a2e-b22d-fc71ae826f14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.360873 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2977934a-8718-4a2e-b22d-fc71ae826f14-kube-api-access-8l7nh" (OuterVolumeSpecName: "kube-api-access-8l7nh") pod "2977934a-8718-4a2e-b22d-fc71ae826f14" (UID: "2977934a-8718-4a2e-b22d-fc71ae826f14"). InnerVolumeSpecName "kube-api-access-8l7nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.420191 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2977934a-8718-4a2e-b22d-fc71ae826f14" (UID: "2977934a-8718-4a2e-b22d-fc71ae826f14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.459309 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.459377 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2977934a-8718-4a2e-b22d-fc71ae826f14-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:02 crc kubenswrapper[4677]: I1203 15:15:02.459395 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l7nh\" (UniqueName: \"kubernetes.io/projected/2977934a-8718-4a2e-b22d-fc71ae826f14-kube-api-access-8l7nh\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.149873 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hmhnx" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.149922 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hmhnx" event={"ID":"2977934a-8718-4a2e-b22d-fc71ae826f14","Type":"ContainerDied","Data":"afdaac3dd7a201a2813ab68efeae9d83fd560498bf577ea6461fb7b16c1033ab"} Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.150353 4677 scope.go:117] "RemoveContainer" containerID="93f09dc5ec5020d34261c7ad1510ff12829b93ac17374242061d05223c03b9de" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.211200 4677 scope.go:117] "RemoveContainer" containerID="6b4178e265459b57e5a703705af3e1c20b756e6731475bd1b2ee44a58005b477" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.221237 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hmhnx"] Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.235671 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hmhnx"] Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.241557 4677 scope.go:117] "RemoveContainer" containerID="d9992bd0ddc46128e5f8913fde47ea86aee44e3e5ff85295d598755cbf7ab60e" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.556250 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.683545 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aecb6a0-009d-43bb-a67f-61e0c2978610-config-volume\") pod \"4aecb6a0-009d-43bb-a67f-61e0c2978610\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.683684 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhfpr\" (UniqueName: \"kubernetes.io/projected/4aecb6a0-009d-43bb-a67f-61e0c2978610-kube-api-access-rhfpr\") pod \"4aecb6a0-009d-43bb-a67f-61e0c2978610\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.683730 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aecb6a0-009d-43bb-a67f-61e0c2978610-secret-volume\") pod \"4aecb6a0-009d-43bb-a67f-61e0c2978610\" (UID: \"4aecb6a0-009d-43bb-a67f-61e0c2978610\") " Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.684503 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4aecb6a0-009d-43bb-a67f-61e0c2978610-config-volume" (OuterVolumeSpecName: "config-volume") pod "4aecb6a0-009d-43bb-a67f-61e0c2978610" (UID: "4aecb6a0-009d-43bb-a67f-61e0c2978610"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.692273 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aecb6a0-009d-43bb-a67f-61e0c2978610-kube-api-access-rhfpr" (OuterVolumeSpecName: "kube-api-access-rhfpr") pod "4aecb6a0-009d-43bb-a67f-61e0c2978610" (UID: "4aecb6a0-009d-43bb-a67f-61e0c2978610"). InnerVolumeSpecName "kube-api-access-rhfpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.696031 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aecb6a0-009d-43bb-a67f-61e0c2978610-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4aecb6a0-009d-43bb-a67f-61e0c2978610" (UID: "4aecb6a0-009d-43bb-a67f-61e0c2978610"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.788601 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhfpr\" (UniqueName: \"kubernetes.io/projected/4aecb6a0-009d-43bb-a67f-61e0c2978610-kube-api-access-rhfpr\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.788660 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aecb6a0-009d-43bb-a67f-61e0c2978610-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.788681 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aecb6a0-009d-43bb-a67f-61e0c2978610-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.996513 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" path="/var/lib/kubelet/pods/21c1c4d5-0c44-4edc-9271-445194132d6e/volumes" Dec 03 15:15:03 crc kubenswrapper[4677]: I1203 15:15:03.998118 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" path="/var/lib/kubelet/pods/2977934a-8718-4a2e-b22d-fc71ae826f14/volumes" Dec 03 15:15:04 crc kubenswrapper[4677]: I1203 15:15:04.160973 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" Dec 03 15:15:04 crc kubenswrapper[4677]: I1203 15:15:04.160982 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q" event={"ID":"4aecb6a0-009d-43bb-a67f-61e0c2978610","Type":"ContainerDied","Data":"650b81cbc074cb9ceaa7af6c602c9de206f302fde96fd2daf6650cb925799ef9"} Dec 03 15:15:04 crc kubenswrapper[4677]: I1203 15:15:04.161056 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="650b81cbc074cb9ceaa7af6c602c9de206f302fde96fd2daf6650cb925799ef9" Dec 03 15:15:04 crc kubenswrapper[4677]: I1203 15:15:04.629778 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms"] Dec 03 15:15:04 crc kubenswrapper[4677]: I1203 15:15:04.638355 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-2tbms"] Dec 03 15:15:05 crc kubenswrapper[4677]: I1203 15:15:05.995084 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef734bef-9a2b-4622-b8e8-d476cebd82ee" path="/var/lib/kubelet/pods/ef734bef-9a2b-4622-b8e8-d476cebd82ee/volumes" Dec 03 15:15:09 crc kubenswrapper[4677]: I1203 15:15:09.985936 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:15:09 crc kubenswrapper[4677]: E1203 15:15:09.986769 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:15:21 crc kubenswrapper[4677]: I1203 15:15:21.976713 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:15:21 crc kubenswrapper[4677]: E1203 15:15:21.978657 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:15:34 crc kubenswrapper[4677]: I1203 15:15:34.976052 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:15:34 crc kubenswrapper[4677]: E1203 15:15:34.976870 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:15:46 crc kubenswrapper[4677]: I1203 15:15:46.976587 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:15:46 crc kubenswrapper[4677]: E1203 15:15:46.977432 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:15:51 crc kubenswrapper[4677]: I1203 15:15:51.483703 4677 scope.go:117] "RemoveContainer" containerID="325ba1ba74b7328ef77ee4badd32e67dbe7042d8c02ab66efa27598016342674" Dec 03 15:15:58 crc kubenswrapper[4677]: I1203 15:15:58.976300 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:15:58 crc kubenswrapper[4677]: E1203 15:15:58.976986 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:16:13 crc kubenswrapper[4677]: I1203 15:16:13.977085 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:16:13 crc kubenswrapper[4677]: E1203 15:16:13.977681 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:16:26 crc kubenswrapper[4677]: I1203 15:16:26.975733 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:16:26 crc kubenswrapper[4677]: E1203 15:16:26.977319 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:16:38 crc kubenswrapper[4677]: I1203 15:16:38.976547 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:16:38 crc kubenswrapper[4677]: E1203 15:16:38.977412 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:16:53 crc kubenswrapper[4677]: I1203 15:16:53.976317 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:16:53 crc kubenswrapper[4677]: E1203 15:16:53.977242 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:17:08 crc kubenswrapper[4677]: I1203 15:17:08.977165 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:17:08 crc kubenswrapper[4677]: E1203 15:17:08.978323 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:17:19 crc kubenswrapper[4677]: I1203 15:17:19.987532 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:17:19 crc kubenswrapper[4677]: E1203 15:17:19.989240 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:17:31 crc kubenswrapper[4677]: I1203 15:17:31.977101 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:17:31 crc kubenswrapper[4677]: E1203 15:17:31.977966 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.856132 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q76ws"] Dec 03 15:17:37 crc kubenswrapper[4677]: E1203 15:17:37.857775 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="extract-utilities" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.857813 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="extract-utilities" Dec 03 15:17:37 crc kubenswrapper[4677]: E1203 15:17:37.857841 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="extract-content" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.857857 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="extract-content" Dec 03 15:17:37 crc kubenswrapper[4677]: E1203 15:17:37.857894 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aecb6a0-009d-43bb-a67f-61e0c2978610" containerName="collect-profiles" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.857910 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aecb6a0-009d-43bb-a67f-61e0c2978610" containerName="collect-profiles" Dec 03 15:17:37 crc kubenswrapper[4677]: E1203 15:17:37.857979 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="extract-utilities" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.857998 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="extract-utilities" Dec 03 15:17:37 crc kubenswrapper[4677]: E1203 15:17:37.858016 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="registry-server" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.858031 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="registry-server" Dec 03 15:17:37 crc kubenswrapper[4677]: E1203 15:17:37.858065 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="extract-content" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.858079 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="extract-content" Dec 03 15:17:37 crc kubenswrapper[4677]: E1203 15:17:37.858138 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="registry-server" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.858155 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="registry-server" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.858705 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c1c4d5-0c44-4edc-9271-445194132d6e" containerName="registry-server" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.858791 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="2977934a-8718-4a2e-b22d-fc71ae826f14" containerName="registry-server" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.858824 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aecb6a0-009d-43bb-a67f-61e0c2978610" containerName="collect-profiles" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.862425 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.885468 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q76ws"] Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.896653 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-utilities\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.896965 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-catalog-content\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.897050 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5stj9\" (UniqueName: \"kubernetes.io/projected/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-kube-api-access-5stj9\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.998472 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-catalog-content\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.998591 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5stj9\" (UniqueName: \"kubernetes.io/projected/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-kube-api-access-5stj9\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.998740 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-utilities\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.999043 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-catalog-content\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:37 crc kubenswrapper[4677]: I1203 15:17:37.999290 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-utilities\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:38 crc kubenswrapper[4677]: I1203 15:17:38.019783 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5stj9\" (UniqueName: \"kubernetes.io/projected/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-kube-api-access-5stj9\") pod \"redhat-marketplace-q76ws\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:38 crc kubenswrapper[4677]: I1203 15:17:38.208028 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:38 crc kubenswrapper[4677]: I1203 15:17:38.707227 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q76ws"] Dec 03 15:17:38 crc kubenswrapper[4677]: I1203 15:17:38.787021 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q76ws" event={"ID":"93fbb86e-7132-4ded-bc45-ecb0e4d880e3","Type":"ContainerStarted","Data":"1bc7cdbf057aa48defd86b34fde456fe253382e4932eb8d5a3b479def4f4d9c0"} Dec 03 15:17:39 crc kubenswrapper[4677]: I1203 15:17:39.804812 4677 generic.go:334] "Generic (PLEG): container finished" podID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerID="09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91" exitCode=0 Dec 03 15:17:39 crc kubenswrapper[4677]: I1203 15:17:39.805004 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q76ws" event={"ID":"93fbb86e-7132-4ded-bc45-ecb0e4d880e3","Type":"ContainerDied","Data":"09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91"} Dec 03 15:17:40 crc kubenswrapper[4677]: I1203 15:17:40.823445 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q76ws" event={"ID":"93fbb86e-7132-4ded-bc45-ecb0e4d880e3","Type":"ContainerStarted","Data":"f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86"} Dec 03 15:17:41 crc kubenswrapper[4677]: I1203 15:17:41.838645 4677 generic.go:334] "Generic (PLEG): container finished" podID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerID="f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86" exitCode=0 Dec 03 15:17:41 crc kubenswrapper[4677]: I1203 15:17:41.838709 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q76ws" event={"ID":"93fbb86e-7132-4ded-bc45-ecb0e4d880e3","Type":"ContainerDied","Data":"f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86"} Dec 03 15:17:42 crc kubenswrapper[4677]: I1203 15:17:42.852231 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q76ws" event={"ID":"93fbb86e-7132-4ded-bc45-ecb0e4d880e3","Type":"ContainerStarted","Data":"56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7"} Dec 03 15:17:42 crc kubenswrapper[4677]: I1203 15:17:42.873374 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q76ws" podStartSLOduration=3.168590597 podStartE2EDuration="5.873349466s" podCreationTimestamp="2025-12-03 15:17:37 +0000 UTC" firstStartedPulling="2025-12-03 15:17:39.807608175 +0000 UTC m=+5450.553940640" lastFinishedPulling="2025-12-03 15:17:42.512367054 +0000 UTC m=+5453.258699509" observedRunningTime="2025-12-03 15:17:42.87238464 +0000 UTC m=+5453.618717115" watchObservedRunningTime="2025-12-03 15:17:42.873349466 +0000 UTC m=+5453.619681931" Dec 03 15:17:44 crc kubenswrapper[4677]: I1203 15:17:44.977101 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:17:44 crc kubenswrapper[4677]: E1203 15:17:44.977764 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:17:48 crc kubenswrapper[4677]: I1203 15:17:48.208231 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:48 crc kubenswrapper[4677]: I1203 15:17:48.209835 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:48 crc kubenswrapper[4677]: I1203 15:17:48.870766 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:48 crc kubenswrapper[4677]: I1203 15:17:48.976741 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:49 crc kubenswrapper[4677]: I1203 15:17:49.650004 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q76ws"] Dec 03 15:17:50 crc kubenswrapper[4677]: I1203 15:17:50.940796 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q76ws" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="registry-server" containerID="cri-o://56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7" gracePeriod=2 Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.456198 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.605382 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-utilities\") pod \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.605641 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5stj9\" (UniqueName: \"kubernetes.io/projected/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-kube-api-access-5stj9\") pod \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.605872 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-catalog-content\") pod \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\" (UID: \"93fbb86e-7132-4ded-bc45-ecb0e4d880e3\") " Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.606360 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-utilities" (OuterVolumeSpecName: "utilities") pod "93fbb86e-7132-4ded-bc45-ecb0e4d880e3" (UID: "93fbb86e-7132-4ded-bc45-ecb0e4d880e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.608456 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.615575 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-kube-api-access-5stj9" (OuterVolumeSpecName: "kube-api-access-5stj9") pod "93fbb86e-7132-4ded-bc45-ecb0e4d880e3" (UID: "93fbb86e-7132-4ded-bc45-ecb0e4d880e3"). InnerVolumeSpecName "kube-api-access-5stj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.630611 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93fbb86e-7132-4ded-bc45-ecb0e4d880e3" (UID: "93fbb86e-7132-4ded-bc45-ecb0e4d880e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.715632 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5stj9\" (UniqueName: \"kubernetes.io/projected/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-kube-api-access-5stj9\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.716021 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93fbb86e-7132-4ded-bc45-ecb0e4d880e3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.956647 4677 generic.go:334] "Generic (PLEG): container finished" podID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerID="56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7" exitCode=0 Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.956778 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q76ws" event={"ID":"93fbb86e-7132-4ded-bc45-ecb0e4d880e3","Type":"ContainerDied","Data":"56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7"} Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.956811 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q76ws" event={"ID":"93fbb86e-7132-4ded-bc45-ecb0e4d880e3","Type":"ContainerDied","Data":"1bc7cdbf057aa48defd86b34fde456fe253382e4932eb8d5a3b479def4f4d9c0"} Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.956848 4677 scope.go:117] "RemoveContainer" containerID="56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.957329 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q76ws" Dec 03 15:17:51 crc kubenswrapper[4677]: I1203 15:17:51.988418 4677 scope.go:117] "RemoveContainer" containerID="f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86" Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.008202 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q76ws"] Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.019785 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q76ws"] Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.028050 4677 scope.go:117] "RemoveContainer" containerID="09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91" Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.086643 4677 scope.go:117] "RemoveContainer" containerID="56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7" Dec 03 15:17:52 crc kubenswrapper[4677]: E1203 15:17:52.087230 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7\": container with ID starting with 56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7 not found: ID does not exist" containerID="56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7" Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.087357 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7"} err="failed to get container status \"56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7\": rpc error: code = NotFound desc = could not find container \"56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7\": container with ID starting with 56300ba79c5921626424c97030e3af21d85407a8bdb73e61e346371142a12ad7 not found: ID does not exist" Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.087394 4677 scope.go:117] "RemoveContainer" containerID="f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86" Dec 03 15:17:52 crc kubenswrapper[4677]: E1203 15:17:52.087971 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86\": container with ID starting with f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86 not found: ID does not exist" containerID="f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86" Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.088019 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86"} err="failed to get container status \"f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86\": rpc error: code = NotFound desc = could not find container \"f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86\": container with ID starting with f3f4b1699b60daaa3cebaabd8eca075d2542550308e423aa6a9c75fbae09be86 not found: ID does not exist" Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.088051 4677 scope.go:117] "RemoveContainer" containerID="09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91" Dec 03 15:17:52 crc kubenswrapper[4677]: E1203 15:17:52.088416 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91\": container with ID starting with 09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91 not found: ID does not exist" containerID="09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91" Dec 03 15:17:52 crc kubenswrapper[4677]: I1203 15:17:52.088470 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91"} err="failed to get container status \"09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91\": rpc error: code = NotFound desc = could not find container \"09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91\": container with ID starting with 09dc67f996644543900980bab825a66f0f76b839d3542d624e312177c3469e91 not found: ID does not exist" Dec 03 15:17:53 crc kubenswrapper[4677]: I1203 15:17:53.999371 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" path="/var/lib/kubelet/pods/93fbb86e-7132-4ded-bc45-ecb0e4d880e3/volumes" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.386288 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9svmr"] Dec 03 15:17:55 crc kubenswrapper[4677]: E1203 15:17:55.387219 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="extract-content" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.387245 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="extract-content" Dec 03 15:17:55 crc kubenswrapper[4677]: E1203 15:17:55.387282 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="extract-utilities" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.387295 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="extract-utilities" Dec 03 15:17:55 crc kubenswrapper[4677]: E1203 15:17:55.387328 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="registry-server" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.387342 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="registry-server" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.387776 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="93fbb86e-7132-4ded-bc45-ecb0e4d880e3" containerName="registry-server" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.390401 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.400578 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9svmr"] Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.519666 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcdbb\" (UniqueName: \"kubernetes.io/projected/a40e377a-7eeb-41ec-89e6-5ce249b21391-kube-api-access-qcdbb\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.519766 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-utilities\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.519823 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-catalog-content\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.621799 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcdbb\" (UniqueName: \"kubernetes.io/projected/a40e377a-7eeb-41ec-89e6-5ce249b21391-kube-api-access-qcdbb\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.621929 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-utilities\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.622026 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-catalog-content\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.622694 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-catalog-content\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.623019 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-utilities\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.651916 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcdbb\" (UniqueName: \"kubernetes.io/projected/a40e377a-7eeb-41ec-89e6-5ce249b21391-kube-api-access-qcdbb\") pod \"certified-operators-9svmr\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:55 crc kubenswrapper[4677]: I1203 15:17:55.726628 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:17:56 crc kubenswrapper[4677]: I1203 15:17:56.321109 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9svmr"] Dec 03 15:17:56 crc kubenswrapper[4677]: I1203 15:17:56.976884 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:17:56 crc kubenswrapper[4677]: E1203 15:17:56.977846 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:17:57 crc kubenswrapper[4677]: I1203 15:17:57.039772 4677 generic.go:334] "Generic (PLEG): container finished" podID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerID="38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609" exitCode=0 Dec 03 15:17:57 crc kubenswrapper[4677]: I1203 15:17:57.039818 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9svmr" event={"ID":"a40e377a-7eeb-41ec-89e6-5ce249b21391","Type":"ContainerDied","Data":"38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609"} Dec 03 15:17:57 crc kubenswrapper[4677]: I1203 15:17:57.039850 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9svmr" event={"ID":"a40e377a-7eeb-41ec-89e6-5ce249b21391","Type":"ContainerStarted","Data":"40e691abf11783174b44a217097967cc70f0bc87ce39ac305d005868a4699df9"} Dec 03 15:17:58 crc kubenswrapper[4677]: I1203 15:17:58.051081 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9svmr" event={"ID":"a40e377a-7eeb-41ec-89e6-5ce249b21391","Type":"ContainerStarted","Data":"013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0"} Dec 03 15:17:59 crc kubenswrapper[4677]: I1203 15:17:59.063335 4677 generic.go:334] "Generic (PLEG): container finished" podID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerID="013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0" exitCode=0 Dec 03 15:17:59 crc kubenswrapper[4677]: I1203 15:17:59.063430 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9svmr" event={"ID":"a40e377a-7eeb-41ec-89e6-5ce249b21391","Type":"ContainerDied","Data":"013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0"} Dec 03 15:18:00 crc kubenswrapper[4677]: I1203 15:18:00.077527 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9svmr" event={"ID":"a40e377a-7eeb-41ec-89e6-5ce249b21391","Type":"ContainerStarted","Data":"9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40"} Dec 03 15:18:00 crc kubenswrapper[4677]: I1203 15:18:00.100394 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9svmr" podStartSLOduration=2.702274715 podStartE2EDuration="5.10037319s" podCreationTimestamp="2025-12-03 15:17:55 +0000 UTC" firstStartedPulling="2025-12-03 15:17:57.042614777 +0000 UTC m=+5467.788947232" lastFinishedPulling="2025-12-03 15:17:59.440713252 +0000 UTC m=+5470.187045707" observedRunningTime="2025-12-03 15:18:00.098665903 +0000 UTC m=+5470.844998368" watchObservedRunningTime="2025-12-03 15:18:00.10037319 +0000 UTC m=+5470.846705655" Dec 03 15:18:05 crc kubenswrapper[4677]: I1203 15:18:05.727807 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:18:05 crc kubenswrapper[4677]: I1203 15:18:05.728417 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:18:05 crc kubenswrapper[4677]: I1203 15:18:05.791794 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:18:06 crc kubenswrapper[4677]: I1203 15:18:06.233731 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:18:06 crc kubenswrapper[4677]: I1203 15:18:06.306664 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9svmr"] Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.165118 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9svmr" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="registry-server" containerID="cri-o://9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40" gracePeriod=2 Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.720399 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.835535 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-utilities\") pod \"a40e377a-7eeb-41ec-89e6-5ce249b21391\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.835660 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcdbb\" (UniqueName: \"kubernetes.io/projected/a40e377a-7eeb-41ec-89e6-5ce249b21391-kube-api-access-qcdbb\") pod \"a40e377a-7eeb-41ec-89e6-5ce249b21391\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.835703 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-catalog-content\") pod \"a40e377a-7eeb-41ec-89e6-5ce249b21391\" (UID: \"a40e377a-7eeb-41ec-89e6-5ce249b21391\") " Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.836620 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-utilities" (OuterVolumeSpecName: "utilities") pod "a40e377a-7eeb-41ec-89e6-5ce249b21391" (UID: "a40e377a-7eeb-41ec-89e6-5ce249b21391"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.837818 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.843342 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40e377a-7eeb-41ec-89e6-5ce249b21391-kube-api-access-qcdbb" (OuterVolumeSpecName: "kube-api-access-qcdbb") pod "a40e377a-7eeb-41ec-89e6-5ce249b21391" (UID: "a40e377a-7eeb-41ec-89e6-5ce249b21391"). InnerVolumeSpecName "kube-api-access-qcdbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:18:08 crc kubenswrapper[4677]: I1203 15:18:08.939860 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcdbb\" (UniqueName: \"kubernetes.io/projected/a40e377a-7eeb-41ec-89e6-5ce249b21391-kube-api-access-qcdbb\") on node \"crc\" DevicePath \"\"" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.182704 4677 generic.go:334] "Generic (PLEG): container finished" podID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerID="9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40" exitCode=0 Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.182760 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9svmr" event={"ID":"a40e377a-7eeb-41ec-89e6-5ce249b21391","Type":"ContainerDied","Data":"9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40"} Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.182790 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9svmr" event={"ID":"a40e377a-7eeb-41ec-89e6-5ce249b21391","Type":"ContainerDied","Data":"40e691abf11783174b44a217097967cc70f0bc87ce39ac305d005868a4699df9"} Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.182809 4677 scope.go:117] "RemoveContainer" containerID="9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.182899 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9svmr" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.215924 4677 scope.go:117] "RemoveContainer" containerID="013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.254516 4677 scope.go:117] "RemoveContainer" containerID="38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.324267 4677 scope.go:117] "RemoveContainer" containerID="9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40" Dec 03 15:18:09 crc kubenswrapper[4677]: E1203 15:18:09.325038 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40\": container with ID starting with 9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40 not found: ID does not exist" containerID="9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.325096 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40"} err="failed to get container status \"9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40\": rpc error: code = NotFound desc = could not find container \"9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40\": container with ID starting with 9a244c8fa1f866610d81e09a0619ae6126b5fd42b1b8ba9be587a8405bc88d40 not found: ID does not exist" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.325134 4677 scope.go:117] "RemoveContainer" containerID="013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0" Dec 03 15:18:09 crc kubenswrapper[4677]: E1203 15:18:09.325570 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0\": container with ID starting with 013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0 not found: ID does not exist" containerID="013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.325616 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0"} err="failed to get container status \"013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0\": rpc error: code = NotFound desc = could not find container \"013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0\": container with ID starting with 013ea4a514d6e28f70d9b4641f8e952a37eb4b62dc492481ffbb10d0615059f0 not found: ID does not exist" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.325651 4677 scope.go:117] "RemoveContainer" containerID="38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609" Dec 03 15:18:09 crc kubenswrapper[4677]: E1203 15:18:09.326364 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609\": container with ID starting with 38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609 not found: ID does not exist" containerID="38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.326469 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609"} err="failed to get container status \"38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609\": rpc error: code = NotFound desc = could not find container \"38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609\": container with ID starting with 38cefdc22874d90a433b4b6fdf601e8dd12040fd574380dc75a0ea73dc434609 not found: ID does not exist" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.774074 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a40e377a-7eeb-41ec-89e6-5ce249b21391" (UID: "a40e377a-7eeb-41ec-89e6-5ce249b21391"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.821603 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9svmr"] Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.832129 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9svmr"] Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.857884 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40e377a-7eeb-41ec-89e6-5ce249b21391-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:18:09 crc kubenswrapper[4677]: I1203 15:18:09.991291 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" path="/var/lib/kubelet/pods/a40e377a-7eeb-41ec-89e6-5ce249b21391/volumes" Dec 03 15:18:11 crc kubenswrapper[4677]: I1203 15:18:11.976368 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:18:13 crc kubenswrapper[4677]: I1203 15:18:13.229763 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"0a6628872426e3d589e9892d195510ec33e8b28f440868d8453c8253b4b0baa2"} Dec 03 15:19:49 crc kubenswrapper[4677]: I1203 15:19:49.741493 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-9cc645f55-v5xrn" podUID="5da71451-5801-4ce9-b177-cd847b635e24" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 03 15:20:38 crc kubenswrapper[4677]: I1203 15:20:38.437549 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:20:38 crc kubenswrapper[4677]: I1203 15:20:38.438226 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:21:08 crc kubenswrapper[4677]: I1203 15:21:08.437567 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:21:08 crc kubenswrapper[4677]: I1203 15:21:08.438267 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:21:38 crc kubenswrapper[4677]: I1203 15:21:38.442939 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:21:38 crc kubenswrapper[4677]: I1203 15:21:38.443649 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:21:38 crc kubenswrapper[4677]: I1203 15:21:38.443703 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:21:38 crc kubenswrapper[4677]: I1203 15:21:38.445212 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a6628872426e3d589e9892d195510ec33e8b28f440868d8453c8253b4b0baa2"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:21:38 crc kubenswrapper[4677]: I1203 15:21:38.445300 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://0a6628872426e3d589e9892d195510ec33e8b28f440868d8453c8253b4b0baa2" gracePeriod=600 Dec 03 15:21:39 crc kubenswrapper[4677]: I1203 15:21:39.421286 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="0a6628872426e3d589e9892d195510ec33e8b28f440868d8453c8253b4b0baa2" exitCode=0 Dec 03 15:21:39 crc kubenswrapper[4677]: I1203 15:21:39.421349 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"0a6628872426e3d589e9892d195510ec33e8b28f440868d8453c8253b4b0baa2"} Dec 03 15:21:39 crc kubenswrapper[4677]: I1203 15:21:39.421796 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b"} Dec 03 15:21:39 crc kubenswrapper[4677]: I1203 15:21:39.421825 4677 scope.go:117] "RemoveContainer" containerID="4afcf9f3e39deaba9c79b43c9aef823aa6d57c0a795bca2e898e3fe96d25a9ce" Dec 03 15:23:38 crc kubenswrapper[4677]: I1203 15:23:38.437493 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:23:38 crc kubenswrapper[4677]: I1203 15:23:38.438185 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:24:08 crc kubenswrapper[4677]: I1203 15:24:08.437707 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:24:08 crc kubenswrapper[4677]: I1203 15:24:08.438290 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:24:38 crc kubenswrapper[4677]: I1203 15:24:38.437100 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:24:38 crc kubenswrapper[4677]: I1203 15:24:38.437709 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:24:38 crc kubenswrapper[4677]: I1203 15:24:38.437810 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:24:38 crc kubenswrapper[4677]: I1203 15:24:38.438823 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:24:38 crc kubenswrapper[4677]: I1203 15:24:38.438902 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" gracePeriod=600 Dec 03 15:24:38 crc kubenswrapper[4677]: E1203 15:24:38.570992 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:24:39 crc kubenswrapper[4677]: I1203 15:24:39.282018 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" exitCode=0 Dec 03 15:24:39 crc kubenswrapper[4677]: I1203 15:24:39.282061 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b"} Dec 03 15:24:39 crc kubenswrapper[4677]: I1203 15:24:39.282228 4677 scope.go:117] "RemoveContainer" containerID="0a6628872426e3d589e9892d195510ec33e8b28f440868d8453c8253b4b0baa2" Dec 03 15:24:39 crc kubenswrapper[4677]: I1203 15:24:39.282875 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:24:39 crc kubenswrapper[4677]: E1203 15:24:39.283290 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:24:50 crc kubenswrapper[4677]: I1203 15:24:50.003718 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:24:50 crc kubenswrapper[4677]: E1203 15:24:50.005813 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:25:03 crc kubenswrapper[4677]: I1203 15:25:03.976052 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:25:03 crc kubenswrapper[4677]: E1203 15:25:03.977195 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.523649 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gf7l9"] Dec 03 15:25:09 crc kubenswrapper[4677]: E1203 15:25:09.524669 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="registry-server" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.524688 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="registry-server" Dec 03 15:25:09 crc kubenswrapper[4677]: E1203 15:25:09.524706 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="extract-content" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.524712 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="extract-content" Dec 03 15:25:09 crc kubenswrapper[4677]: E1203 15:25:09.524746 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="extract-utilities" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.524753 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="extract-utilities" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.530406 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a40e377a-7eeb-41ec-89e6-5ce249b21391" containerName="registry-server" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.535619 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.536519 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gf7l9"] Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.679622 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8cq6\" (UniqueName: \"kubernetes.io/projected/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-kube-api-access-g8cq6\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.679872 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-catalog-content\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.680224 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-utilities\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.782731 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8cq6\" (UniqueName: \"kubernetes.io/projected/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-kube-api-access-g8cq6\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.782915 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-catalog-content\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.783049 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-utilities\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.783635 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-utilities\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.783652 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-catalog-content\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.811264 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8cq6\" (UniqueName: \"kubernetes.io/projected/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-kube-api-access-g8cq6\") pod \"redhat-operators-gf7l9\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:09 crc kubenswrapper[4677]: I1203 15:25:09.870068 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:10 crc kubenswrapper[4677]: I1203 15:25:10.377309 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gf7l9"] Dec 03 15:25:10 crc kubenswrapper[4677]: I1203 15:25:10.612419 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gf7l9" event={"ID":"6e6f1ceb-9aac-4614-aa3b-6981de139c3f","Type":"ContainerStarted","Data":"e178b4683a21c7367287b0663ce7ee928698ea307d86b8cf7a508acfd0f3e92d"} Dec 03 15:25:11 crc kubenswrapper[4677]: I1203 15:25:11.623033 4677 generic.go:334] "Generic (PLEG): container finished" podID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerID="4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec" exitCode=0 Dec 03 15:25:11 crc kubenswrapper[4677]: I1203 15:25:11.623098 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gf7l9" event={"ID":"6e6f1ceb-9aac-4614-aa3b-6981de139c3f","Type":"ContainerDied","Data":"4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec"} Dec 03 15:25:11 crc kubenswrapper[4677]: I1203 15:25:11.627168 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:25:12 crc kubenswrapper[4677]: I1203 15:25:12.634418 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gf7l9" event={"ID":"6e6f1ceb-9aac-4614-aa3b-6981de139c3f","Type":"ContainerStarted","Data":"8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55"} Dec 03 15:25:14 crc kubenswrapper[4677]: I1203 15:25:14.977176 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:25:14 crc kubenswrapper[4677]: E1203 15:25:14.978620 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:25:15 crc kubenswrapper[4677]: I1203 15:25:15.670653 4677 generic.go:334] "Generic (PLEG): container finished" podID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerID="8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55" exitCode=0 Dec 03 15:25:15 crc kubenswrapper[4677]: I1203 15:25:15.670700 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gf7l9" event={"ID":"6e6f1ceb-9aac-4614-aa3b-6981de139c3f","Type":"ContainerDied","Data":"8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55"} Dec 03 15:25:17 crc kubenswrapper[4677]: I1203 15:25:17.693323 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gf7l9" event={"ID":"6e6f1ceb-9aac-4614-aa3b-6981de139c3f","Type":"ContainerStarted","Data":"988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753"} Dec 03 15:25:17 crc kubenswrapper[4677]: I1203 15:25:17.712697 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gf7l9" podStartSLOduration=3.059636877 podStartE2EDuration="8.712680855s" podCreationTimestamp="2025-12-03 15:25:09 +0000 UTC" firstStartedPulling="2025-12-03 15:25:11.626963563 +0000 UTC m=+5902.373296018" lastFinishedPulling="2025-12-03 15:25:17.280007541 +0000 UTC m=+5908.026339996" observedRunningTime="2025-12-03 15:25:17.71104457 +0000 UTC m=+5908.457377045" watchObservedRunningTime="2025-12-03 15:25:17.712680855 +0000 UTC m=+5908.459013310" Dec 03 15:25:19 crc kubenswrapper[4677]: I1203 15:25:19.870490 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:19 crc kubenswrapper[4677]: I1203 15:25:19.870813 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:20 crc kubenswrapper[4677]: I1203 15:25:20.918024 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gf7l9" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="registry-server" probeResult="failure" output=< Dec 03 15:25:20 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 15:25:20 crc kubenswrapper[4677]: > Dec 03 15:25:25 crc kubenswrapper[4677]: I1203 15:25:25.976436 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:25:25 crc kubenswrapper[4677]: E1203 15:25:25.977138 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:25:29 crc kubenswrapper[4677]: I1203 15:25:29.922438 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:29 crc kubenswrapper[4677]: I1203 15:25:29.973552 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:30 crc kubenswrapper[4677]: I1203 15:25:30.160386 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gf7l9"] Dec 03 15:25:31 crc kubenswrapper[4677]: I1203 15:25:31.842697 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gf7l9" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="registry-server" containerID="cri-o://988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753" gracePeriod=2 Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.322275 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.448055 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-utilities\") pod \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.448342 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8cq6\" (UniqueName: \"kubernetes.io/projected/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-kube-api-access-g8cq6\") pod \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.448386 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-catalog-content\") pod \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\" (UID: \"6e6f1ceb-9aac-4614-aa3b-6981de139c3f\") " Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.448816 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-utilities" (OuterVolumeSpecName: "utilities") pod "6e6f1ceb-9aac-4614-aa3b-6981de139c3f" (UID: "6e6f1ceb-9aac-4614-aa3b-6981de139c3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.449084 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.454695 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-kube-api-access-g8cq6" (OuterVolumeSpecName: "kube-api-access-g8cq6") pod "6e6f1ceb-9aac-4614-aa3b-6981de139c3f" (UID: "6e6f1ceb-9aac-4614-aa3b-6981de139c3f"). InnerVolumeSpecName "kube-api-access-g8cq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.551274 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8cq6\" (UniqueName: \"kubernetes.io/projected/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-kube-api-access-g8cq6\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.566772 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e6f1ceb-9aac-4614-aa3b-6981de139c3f" (UID: "6e6f1ceb-9aac-4614-aa3b-6981de139c3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.653153 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6f1ceb-9aac-4614-aa3b-6981de139c3f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.858790 4677 generic.go:334] "Generic (PLEG): container finished" podID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerID="988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753" exitCode=0 Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.858846 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gf7l9" event={"ID":"6e6f1ceb-9aac-4614-aa3b-6981de139c3f","Type":"ContainerDied","Data":"988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753"} Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.858917 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gf7l9" event={"ID":"6e6f1ceb-9aac-4614-aa3b-6981de139c3f","Type":"ContainerDied","Data":"e178b4683a21c7367287b0663ce7ee928698ea307d86b8cf7a508acfd0f3e92d"} Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.858867 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gf7l9" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.858967 4677 scope.go:117] "RemoveContainer" containerID="988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.880122 4677 scope.go:117] "RemoveContainer" containerID="8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.898461 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gf7l9"] Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.909691 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gf7l9"] Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.924355 4677 scope.go:117] "RemoveContainer" containerID="4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.976682 4677 scope.go:117] "RemoveContainer" containerID="988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753" Dec 03 15:25:32 crc kubenswrapper[4677]: E1203 15:25:32.977222 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753\": container with ID starting with 988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753 not found: ID does not exist" containerID="988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.977285 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753"} err="failed to get container status \"988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753\": rpc error: code = NotFound desc = could not find container \"988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753\": container with ID starting with 988fd9ce011446ab33ecf0a87c975bc8e5807cd9dafb549f855445f230ce0753 not found: ID does not exist" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.977307 4677 scope.go:117] "RemoveContainer" containerID="8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55" Dec 03 15:25:32 crc kubenswrapper[4677]: E1203 15:25:32.977789 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55\": container with ID starting with 8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55 not found: ID does not exist" containerID="8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.977842 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55"} err="failed to get container status \"8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55\": rpc error: code = NotFound desc = could not find container \"8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55\": container with ID starting with 8505ef764bccc137101e7eeabd8d6d3880673dcf82e35dec8a678c09fcbb4b55 not found: ID does not exist" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.977877 4677 scope.go:117] "RemoveContainer" containerID="4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec" Dec 03 15:25:32 crc kubenswrapper[4677]: E1203 15:25:32.978311 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec\": container with ID starting with 4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec not found: ID does not exist" containerID="4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec" Dec 03 15:25:32 crc kubenswrapper[4677]: I1203 15:25:32.978345 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec"} err="failed to get container status \"4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec\": rpc error: code = NotFound desc = could not find container \"4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec\": container with ID starting with 4adab51f39ac9f94c70b1b2601d5ef611dd6de5272d98c43ac1fbb9ea1603bec not found: ID does not exist" Dec 03 15:25:33 crc kubenswrapper[4677]: I1203 15:25:33.990842 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" path="/var/lib/kubelet/pods/6e6f1ceb-9aac-4614-aa3b-6981de139c3f/volumes" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.568000 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nk7cr"] Dec 03 15:25:35 crc kubenswrapper[4677]: E1203 15:25:35.568747 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="extract-content" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.568764 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="extract-content" Dec 03 15:25:35 crc kubenswrapper[4677]: E1203 15:25:35.568786 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="extract-utilities" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.568793 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="extract-utilities" Dec 03 15:25:35 crc kubenswrapper[4677]: E1203 15:25:35.568821 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="registry-server" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.568828 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="registry-server" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.569044 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e6f1ceb-9aac-4614-aa3b-6981de139c3f" containerName="registry-server" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.572809 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.581925 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nk7cr"] Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.724115 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-utilities\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.724181 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdgtz\" (UniqueName: \"kubernetes.io/projected/e0c1c868-a593-4b37-a822-6f359627431e-kube-api-access-gdgtz\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.724217 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-catalog-content\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.826416 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-catalog-content\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.826672 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-utilities\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.826726 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdgtz\" (UniqueName: \"kubernetes.io/projected/e0c1c868-a593-4b37-a822-6f359627431e-kube-api-access-gdgtz\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.827639 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-catalog-content\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.827665 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-utilities\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.854814 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdgtz\" (UniqueName: \"kubernetes.io/projected/e0c1c868-a593-4b37-a822-6f359627431e-kube-api-access-gdgtz\") pod \"community-operators-nk7cr\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:35 crc kubenswrapper[4677]: I1203 15:25:35.910582 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:36 crc kubenswrapper[4677]: I1203 15:25:36.507036 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nk7cr"] Dec 03 15:25:36 crc kubenswrapper[4677]: I1203 15:25:36.912998 4677 generic.go:334] "Generic (PLEG): container finished" podID="e0c1c868-a593-4b37-a822-6f359627431e" containerID="3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6" exitCode=0 Dec 03 15:25:36 crc kubenswrapper[4677]: I1203 15:25:36.913097 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nk7cr" event={"ID":"e0c1c868-a593-4b37-a822-6f359627431e","Type":"ContainerDied","Data":"3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6"} Dec 03 15:25:36 crc kubenswrapper[4677]: I1203 15:25:36.913290 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nk7cr" event={"ID":"e0c1c868-a593-4b37-a822-6f359627431e","Type":"ContainerStarted","Data":"28022d52b4483554679f412d32a63a1b75c3ffe7aa20efe8230cb41a7eb4272f"} Dec 03 15:25:38 crc kubenswrapper[4677]: I1203 15:25:38.936166 4677 generic.go:334] "Generic (PLEG): container finished" podID="e0c1c868-a593-4b37-a822-6f359627431e" containerID="f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f" exitCode=0 Dec 03 15:25:38 crc kubenswrapper[4677]: I1203 15:25:38.936356 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nk7cr" event={"ID":"e0c1c868-a593-4b37-a822-6f359627431e","Type":"ContainerDied","Data":"f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f"} Dec 03 15:25:38 crc kubenswrapper[4677]: I1203 15:25:38.975913 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:25:38 crc kubenswrapper[4677]: E1203 15:25:38.976606 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:25:39 crc kubenswrapper[4677]: I1203 15:25:39.951516 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nk7cr" event={"ID":"e0c1c868-a593-4b37-a822-6f359627431e","Type":"ContainerStarted","Data":"769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26"} Dec 03 15:25:39 crc kubenswrapper[4677]: I1203 15:25:39.981871 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nk7cr" podStartSLOduration=2.525367026 podStartE2EDuration="4.981851183s" podCreationTimestamp="2025-12-03 15:25:35 +0000 UTC" firstStartedPulling="2025-12-03 15:25:36.915879306 +0000 UTC m=+5927.662211761" lastFinishedPulling="2025-12-03 15:25:39.372363463 +0000 UTC m=+5930.118695918" observedRunningTime="2025-12-03 15:25:39.969061924 +0000 UTC m=+5930.715394399" watchObservedRunningTime="2025-12-03 15:25:39.981851183 +0000 UTC m=+5930.728183638" Dec 03 15:25:45 crc kubenswrapper[4677]: I1203 15:25:45.911346 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:45 crc kubenswrapper[4677]: I1203 15:25:45.913189 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:45 crc kubenswrapper[4677]: I1203 15:25:45.968151 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:46 crc kubenswrapper[4677]: I1203 15:25:46.086660 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:46 crc kubenswrapper[4677]: I1203 15:25:46.210455 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nk7cr"] Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.050926 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nk7cr" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="registry-server" containerID="cri-o://769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26" gracePeriod=2 Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.581889 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.724560 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-utilities\") pod \"e0c1c868-a593-4b37-a822-6f359627431e\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.724902 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-catalog-content\") pod \"e0c1c868-a593-4b37-a822-6f359627431e\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.725215 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdgtz\" (UniqueName: \"kubernetes.io/projected/e0c1c868-a593-4b37-a822-6f359627431e-kube-api-access-gdgtz\") pod \"e0c1c868-a593-4b37-a822-6f359627431e\" (UID: \"e0c1c868-a593-4b37-a822-6f359627431e\") " Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.725800 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-utilities" (OuterVolumeSpecName: "utilities") pod "e0c1c868-a593-4b37-a822-6f359627431e" (UID: "e0c1c868-a593-4b37-a822-6f359627431e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.726165 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.735477 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0c1c868-a593-4b37-a822-6f359627431e-kube-api-access-gdgtz" (OuterVolumeSpecName: "kube-api-access-gdgtz") pod "e0c1c868-a593-4b37-a822-6f359627431e" (UID: "e0c1c868-a593-4b37-a822-6f359627431e"). InnerVolumeSpecName "kube-api-access-gdgtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.828208 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdgtz\" (UniqueName: \"kubernetes.io/projected/e0c1c868-a593-4b37-a822-6f359627431e-kube-api-access-gdgtz\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.834758 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0c1c868-a593-4b37-a822-6f359627431e" (UID: "e0c1c868-a593-4b37-a822-6f359627431e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:25:48 crc kubenswrapper[4677]: I1203 15:25:48.930494 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c1c868-a593-4b37-a822-6f359627431e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.063008 4677 generic.go:334] "Generic (PLEG): container finished" podID="e0c1c868-a593-4b37-a822-6f359627431e" containerID="769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26" exitCode=0 Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.063065 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nk7cr" event={"ID":"e0c1c868-a593-4b37-a822-6f359627431e","Type":"ContainerDied","Data":"769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26"} Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.063080 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nk7cr" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.063106 4677 scope.go:117] "RemoveContainer" containerID="769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.063094 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nk7cr" event={"ID":"e0c1c868-a593-4b37-a822-6f359627431e","Type":"ContainerDied","Data":"28022d52b4483554679f412d32a63a1b75c3ffe7aa20efe8230cb41a7eb4272f"} Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.095804 4677 scope.go:117] "RemoveContainer" containerID="f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.107001 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nk7cr"] Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.118129 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nk7cr"] Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.142510 4677 scope.go:117] "RemoveContainer" containerID="3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.183124 4677 scope.go:117] "RemoveContainer" containerID="769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26" Dec 03 15:25:49 crc kubenswrapper[4677]: E1203 15:25:49.183754 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26\": container with ID starting with 769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26 not found: ID does not exist" containerID="769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.183796 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26"} err="failed to get container status \"769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26\": rpc error: code = NotFound desc = could not find container \"769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26\": container with ID starting with 769ea8399b9ce72fe613a72dab4b7a6bfb9ed477944ec8917c3e4af160307f26 not found: ID does not exist" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.183830 4677 scope.go:117] "RemoveContainer" containerID="f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f" Dec 03 15:25:49 crc kubenswrapper[4677]: E1203 15:25:49.184080 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f\": container with ID starting with f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f not found: ID does not exist" containerID="f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.184101 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f"} err="failed to get container status \"f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f\": rpc error: code = NotFound desc = could not find container \"f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f\": container with ID starting with f86b4f51dc48c1cdaa08af5fbf6657883df35557badd43836043087c0362877f not found: ID does not exist" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.184151 4677 scope.go:117] "RemoveContainer" containerID="3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6" Dec 03 15:25:49 crc kubenswrapper[4677]: E1203 15:25:49.184432 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6\": container with ID starting with 3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6 not found: ID does not exist" containerID="3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.184466 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6"} err="failed to get container status \"3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6\": rpc error: code = NotFound desc = could not find container \"3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6\": container with ID starting with 3920f71dee170e0af158c307dfffe63be5ee191051fd1975c97d538951e6f0d6 not found: ID does not exist" Dec 03 15:25:49 crc kubenswrapper[4677]: I1203 15:25:49.991264 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0c1c868-a593-4b37-a822-6f359627431e" path="/var/lib/kubelet/pods/e0c1c868-a593-4b37-a822-6f359627431e/volumes" Dec 03 15:25:52 crc kubenswrapper[4677]: I1203 15:25:52.976758 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:25:52 crc kubenswrapper[4677]: E1203 15:25:52.977538 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:26:03 crc kubenswrapper[4677]: I1203 15:26:03.976461 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:26:03 crc kubenswrapper[4677]: E1203 15:26:03.977362 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:26:16 crc kubenswrapper[4677]: I1203 15:26:16.976039 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:26:16 crc kubenswrapper[4677]: E1203 15:26:16.977113 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:26:31 crc kubenswrapper[4677]: I1203 15:26:31.976536 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:26:31 crc kubenswrapper[4677]: E1203 15:26:31.977558 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:26:42 crc kubenswrapper[4677]: I1203 15:26:42.977045 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:26:42 crc kubenswrapper[4677]: E1203 15:26:42.978756 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:26:53 crc kubenswrapper[4677]: I1203 15:26:53.977100 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:26:53 crc kubenswrapper[4677]: E1203 15:26:53.978229 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:27:04 crc kubenswrapper[4677]: I1203 15:27:04.977533 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:27:04 crc kubenswrapper[4677]: E1203 15:27:04.978795 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:27:19 crc kubenswrapper[4677]: I1203 15:27:19.975978 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:27:19 crc kubenswrapper[4677]: E1203 15:27:19.977466 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:27:31 crc kubenswrapper[4677]: I1203 15:27:31.976872 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:27:31 crc kubenswrapper[4677]: E1203 15:27:31.977804 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:27:44 crc kubenswrapper[4677]: I1203 15:27:44.976657 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:27:44 crc kubenswrapper[4677]: E1203 15:27:44.977784 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:27:58 crc kubenswrapper[4677]: I1203 15:27:58.976694 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:27:58 crc kubenswrapper[4677]: E1203 15:27:58.977490 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:28:13 crc kubenswrapper[4677]: I1203 15:28:13.976766 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:28:13 crc kubenswrapper[4677]: E1203 15:28:13.977583 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:28:22 crc kubenswrapper[4677]: I1203 15:28:22.815777 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ds8rp"] Dec 03 15:28:22 crc kubenswrapper[4677]: E1203 15:28:22.817428 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="extract-utilities" Dec 03 15:28:22 crc kubenswrapper[4677]: I1203 15:28:22.817479 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="extract-utilities" Dec 03 15:28:22 crc kubenswrapper[4677]: E1203 15:28:22.817544 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="extract-content" Dec 03 15:28:22 crc kubenswrapper[4677]: I1203 15:28:22.817553 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="extract-content" Dec 03 15:28:22 crc kubenswrapper[4677]: E1203 15:28:22.817596 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="registry-server" Dec 03 15:28:22 crc kubenswrapper[4677]: I1203 15:28:22.817605 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="registry-server" Dec 03 15:28:22 crc kubenswrapper[4677]: I1203 15:28:22.819911 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c1c868-a593-4b37-a822-6f359627431e" containerName="registry-server" Dec 03 15:28:22 crc kubenswrapper[4677]: I1203 15:28:22.825678 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:22 crc kubenswrapper[4677]: I1203 15:28:22.833711 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ds8rp"] Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.007657 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-utilities\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.007797 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-catalog-content\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.007891 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpv5b\" (UniqueName: \"kubernetes.io/projected/351fe843-d602-46b5-81ca-2468430d5eb1-kube-api-access-rpv5b\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.110407 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-catalog-content\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.110543 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpv5b\" (UniqueName: \"kubernetes.io/projected/351fe843-d602-46b5-81ca-2468430d5eb1-kube-api-access-rpv5b\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.110725 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-utilities\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.111179 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-catalog-content\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.111499 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-utilities\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.134286 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpv5b\" (UniqueName: \"kubernetes.io/projected/351fe843-d602-46b5-81ca-2468430d5eb1-kube-api-access-rpv5b\") pod \"certified-operators-ds8rp\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.207349 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:23 crc kubenswrapper[4677]: I1203 15:28:23.811758 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ds8rp"] Dec 03 15:28:24 crc kubenswrapper[4677]: I1203 15:28:24.701505 4677 generic.go:334] "Generic (PLEG): container finished" podID="351fe843-d602-46b5-81ca-2468430d5eb1" containerID="9b920a33e2e1c51cc164662986f47150a80164d9bdfd1f25d6920b3bcea3cc6f" exitCode=0 Dec 03 15:28:24 crc kubenswrapper[4677]: I1203 15:28:24.701625 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ds8rp" event={"ID":"351fe843-d602-46b5-81ca-2468430d5eb1","Type":"ContainerDied","Data":"9b920a33e2e1c51cc164662986f47150a80164d9bdfd1f25d6920b3bcea3cc6f"} Dec 03 15:28:24 crc kubenswrapper[4677]: I1203 15:28:24.702996 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ds8rp" event={"ID":"351fe843-d602-46b5-81ca-2468430d5eb1","Type":"ContainerStarted","Data":"4f54e709e0239ae63da5448c062b246c2e39b818e786b874ad8142bff956a1a3"} Dec 03 15:28:24 crc kubenswrapper[4677]: I1203 15:28:24.976578 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:28:24 crc kubenswrapper[4677]: E1203 15:28:24.976876 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:28:26 crc kubenswrapper[4677]: I1203 15:28:26.736466 4677 generic.go:334] "Generic (PLEG): container finished" podID="351fe843-d602-46b5-81ca-2468430d5eb1" containerID="a982b2f065068b44ff9585567bd7361c2865e26b36b8886a49dcfaa0811b3ed1" exitCode=0 Dec 03 15:28:26 crc kubenswrapper[4677]: I1203 15:28:26.738286 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ds8rp" event={"ID":"351fe843-d602-46b5-81ca-2468430d5eb1","Type":"ContainerDied","Data":"a982b2f065068b44ff9585567bd7361c2865e26b36b8886a49dcfaa0811b3ed1"} Dec 03 15:28:28 crc kubenswrapper[4677]: I1203 15:28:28.759995 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ds8rp" event={"ID":"351fe843-d602-46b5-81ca-2468430d5eb1","Type":"ContainerStarted","Data":"e2e8ac7757d7949bab215c2e3c12ee0dcb8487f73c1adef849f4cbe7aa2ccd76"} Dec 03 15:28:28 crc kubenswrapper[4677]: I1203 15:28:28.779815 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ds8rp" podStartSLOduration=3.016855548 podStartE2EDuration="6.779792663s" podCreationTimestamp="2025-12-03 15:28:22 +0000 UTC" firstStartedPulling="2025-12-03 15:28:24.70375088 +0000 UTC m=+6095.450083335" lastFinishedPulling="2025-12-03 15:28:28.466687995 +0000 UTC m=+6099.213020450" observedRunningTime="2025-12-03 15:28:28.775806565 +0000 UTC m=+6099.522139040" watchObservedRunningTime="2025-12-03 15:28:28.779792663 +0000 UTC m=+6099.526125128" Dec 03 15:28:33 crc kubenswrapper[4677]: I1203 15:28:33.208263 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:33 crc kubenswrapper[4677]: I1203 15:28:33.208679 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:33 crc kubenswrapper[4677]: I1203 15:28:33.273295 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:33 crc kubenswrapper[4677]: I1203 15:28:33.866042 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.098054 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-swrmv"] Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.104426 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.128083 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-swrmv"] Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.263158 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-utilities\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.263896 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz7d2\" (UniqueName: \"kubernetes.io/projected/3062d8a6-830f-4daa-bc22-9381e8d00808-kube-api-access-wz7d2\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.264051 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-catalog-content\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.366962 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-utilities\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.367080 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz7d2\" (UniqueName: \"kubernetes.io/projected/3062d8a6-830f-4daa-bc22-9381e8d00808-kube-api-access-wz7d2\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.367117 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-catalog-content\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.370527 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-catalog-content\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.370562 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-utilities\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.390297 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz7d2\" (UniqueName: \"kubernetes.io/projected/3062d8a6-830f-4daa-bc22-9381e8d00808-kube-api-access-wz7d2\") pod \"redhat-marketplace-swrmv\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.447450 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:34 crc kubenswrapper[4677]: I1203 15:28:34.994702 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-swrmv"] Dec 03 15:28:35 crc kubenswrapper[4677]: W1203 15:28:35.009204 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3062d8a6_830f_4daa_bc22_9381e8d00808.slice/crio-ab24905f5897ee5428cc0e91838579eea78550cc61e47f6e27fed5d46ef168f0 WatchSource:0}: Error finding container ab24905f5897ee5428cc0e91838579eea78550cc61e47f6e27fed5d46ef168f0: Status 404 returned error can't find the container with id ab24905f5897ee5428cc0e91838579eea78550cc61e47f6e27fed5d46ef168f0 Dec 03 15:28:35 crc kubenswrapper[4677]: I1203 15:28:35.838910 4677 generic.go:334] "Generic (PLEG): container finished" podID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerID="647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc" exitCode=0 Dec 03 15:28:35 crc kubenswrapper[4677]: I1203 15:28:35.839027 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swrmv" event={"ID":"3062d8a6-830f-4daa-bc22-9381e8d00808","Type":"ContainerDied","Data":"647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc"} Dec 03 15:28:35 crc kubenswrapper[4677]: I1203 15:28:35.839349 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swrmv" event={"ID":"3062d8a6-830f-4daa-bc22-9381e8d00808","Type":"ContainerStarted","Data":"ab24905f5897ee5428cc0e91838579eea78550cc61e47f6e27fed5d46ef168f0"} Dec 03 15:28:36 crc kubenswrapper[4677]: I1203 15:28:36.487235 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ds8rp"] Dec 03 15:28:36 crc kubenswrapper[4677]: I1203 15:28:36.487733 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ds8rp" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="registry-server" containerID="cri-o://e2e8ac7757d7949bab215c2e3c12ee0dcb8487f73c1adef849f4cbe7aa2ccd76" gracePeriod=2 Dec 03 15:28:36 crc kubenswrapper[4677]: I1203 15:28:36.864593 4677 generic.go:334] "Generic (PLEG): container finished" podID="351fe843-d602-46b5-81ca-2468430d5eb1" containerID="e2e8ac7757d7949bab215c2e3c12ee0dcb8487f73c1adef849f4cbe7aa2ccd76" exitCode=0 Dec 03 15:28:36 crc kubenswrapper[4677]: I1203 15:28:36.864635 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ds8rp" event={"ID":"351fe843-d602-46b5-81ca-2468430d5eb1","Type":"ContainerDied","Data":"e2e8ac7757d7949bab215c2e3c12ee0dcb8487f73c1adef849f4cbe7aa2ccd76"} Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.022098 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.127050 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-utilities\") pod \"351fe843-d602-46b5-81ca-2468430d5eb1\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.127119 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-catalog-content\") pod \"351fe843-d602-46b5-81ca-2468430d5eb1\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.127163 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpv5b\" (UniqueName: \"kubernetes.io/projected/351fe843-d602-46b5-81ca-2468430d5eb1-kube-api-access-rpv5b\") pod \"351fe843-d602-46b5-81ca-2468430d5eb1\" (UID: \"351fe843-d602-46b5-81ca-2468430d5eb1\") " Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.128691 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-utilities" (OuterVolumeSpecName: "utilities") pod "351fe843-d602-46b5-81ca-2468430d5eb1" (UID: "351fe843-d602-46b5-81ca-2468430d5eb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.138281 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/351fe843-d602-46b5-81ca-2468430d5eb1-kube-api-access-rpv5b" (OuterVolumeSpecName: "kube-api-access-rpv5b") pod "351fe843-d602-46b5-81ca-2468430d5eb1" (UID: "351fe843-d602-46b5-81ca-2468430d5eb1"). InnerVolumeSpecName "kube-api-access-rpv5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.190122 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "351fe843-d602-46b5-81ca-2468430d5eb1" (UID: "351fe843-d602-46b5-81ca-2468430d5eb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.230563 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.230711 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351fe843-d602-46b5-81ca-2468430d5eb1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.230772 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpv5b\" (UniqueName: \"kubernetes.io/projected/351fe843-d602-46b5-81ca-2468430d5eb1-kube-api-access-rpv5b\") on node \"crc\" DevicePath \"\"" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.879761 4677 generic.go:334] "Generic (PLEG): container finished" podID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerID="6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116" exitCode=0 Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.879844 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swrmv" event={"ID":"3062d8a6-830f-4daa-bc22-9381e8d00808","Type":"ContainerDied","Data":"6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116"} Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.883642 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ds8rp" event={"ID":"351fe843-d602-46b5-81ca-2468430d5eb1","Type":"ContainerDied","Data":"4f54e709e0239ae63da5448c062b246c2e39b818e786b874ad8142bff956a1a3"} Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.883714 4677 scope.go:117] "RemoveContainer" containerID="e2e8ac7757d7949bab215c2e3c12ee0dcb8487f73c1adef849f4cbe7aa2ccd76" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.883756 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ds8rp" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.914823 4677 scope.go:117] "RemoveContainer" containerID="a982b2f065068b44ff9585567bd7361c2865e26b36b8886a49dcfaa0811b3ed1" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.946700 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ds8rp"] Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.957482 4677 scope.go:117] "RemoveContainer" containerID="9b920a33e2e1c51cc164662986f47150a80164d9bdfd1f25d6920b3bcea3cc6f" Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.958648 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ds8rp"] Dec 03 15:28:37 crc kubenswrapper[4677]: I1203 15:28:37.988730 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" path="/var/lib/kubelet/pods/351fe843-d602-46b5-81ca-2468430d5eb1/volumes" Dec 03 15:28:38 crc kubenswrapper[4677]: I1203 15:28:38.896773 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swrmv" event={"ID":"3062d8a6-830f-4daa-bc22-9381e8d00808","Type":"ContainerStarted","Data":"91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075"} Dec 03 15:28:38 crc kubenswrapper[4677]: I1203 15:28:38.916995 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-swrmv" podStartSLOduration=2.446744336 podStartE2EDuration="4.916979758s" podCreationTimestamp="2025-12-03 15:28:34 +0000 UTC" firstStartedPulling="2025-12-03 15:28:35.840996977 +0000 UTC m=+6106.587329442" lastFinishedPulling="2025-12-03 15:28:38.311232389 +0000 UTC m=+6109.057564864" observedRunningTime="2025-12-03 15:28:38.913145882 +0000 UTC m=+6109.659478347" watchObservedRunningTime="2025-12-03 15:28:38.916979758 +0000 UTC m=+6109.663312223" Dec 03 15:28:39 crc kubenswrapper[4677]: I1203 15:28:38.988547 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:28:39 crc kubenswrapper[4677]: E1203 15:28:38.988837 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:28:44 crc kubenswrapper[4677]: I1203 15:28:44.448511 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:44 crc kubenswrapper[4677]: I1203 15:28:44.449752 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:44 crc kubenswrapper[4677]: I1203 15:28:44.538927 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:45 crc kubenswrapper[4677]: I1203 15:28:45.021260 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:45 crc kubenswrapper[4677]: I1203 15:28:45.077678 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-swrmv"] Dec 03 15:28:46 crc kubenswrapper[4677]: I1203 15:28:46.984860 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-swrmv" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="registry-server" containerID="cri-o://91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075" gracePeriod=2 Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.492227 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.669838 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-utilities\") pod \"3062d8a6-830f-4daa-bc22-9381e8d00808\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.669973 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz7d2\" (UniqueName: \"kubernetes.io/projected/3062d8a6-830f-4daa-bc22-9381e8d00808-kube-api-access-wz7d2\") pod \"3062d8a6-830f-4daa-bc22-9381e8d00808\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.670074 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-catalog-content\") pod \"3062d8a6-830f-4daa-bc22-9381e8d00808\" (UID: \"3062d8a6-830f-4daa-bc22-9381e8d00808\") " Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.671351 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-utilities" (OuterVolumeSpecName: "utilities") pod "3062d8a6-830f-4daa-bc22-9381e8d00808" (UID: "3062d8a6-830f-4daa-bc22-9381e8d00808"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.677487 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3062d8a6-830f-4daa-bc22-9381e8d00808-kube-api-access-wz7d2" (OuterVolumeSpecName: "kube-api-access-wz7d2") pod "3062d8a6-830f-4daa-bc22-9381e8d00808" (UID: "3062d8a6-830f-4daa-bc22-9381e8d00808"). InnerVolumeSpecName "kube-api-access-wz7d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.691034 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3062d8a6-830f-4daa-bc22-9381e8d00808" (UID: "3062d8a6-830f-4daa-bc22-9381e8d00808"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.773215 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.773256 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3062d8a6-830f-4daa-bc22-9381e8d00808-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:28:47 crc kubenswrapper[4677]: I1203 15:28:47.773269 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz7d2\" (UniqueName: \"kubernetes.io/projected/3062d8a6-830f-4daa-bc22-9381e8d00808-kube-api-access-wz7d2\") on node \"crc\" DevicePath \"\"" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.000070 4677 generic.go:334] "Generic (PLEG): container finished" podID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerID="91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075" exitCode=0 Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.000123 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swrmv" event={"ID":"3062d8a6-830f-4daa-bc22-9381e8d00808","Type":"ContainerDied","Data":"91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075"} Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.000156 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swrmv" event={"ID":"3062d8a6-830f-4daa-bc22-9381e8d00808","Type":"ContainerDied","Data":"ab24905f5897ee5428cc0e91838579eea78550cc61e47f6e27fed5d46ef168f0"} Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.000179 4677 scope.go:117] "RemoveContainer" containerID="91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.000243 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swrmv" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.080867 4677 scope.go:117] "RemoveContainer" containerID="6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.090226 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-swrmv"] Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.109233 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-swrmv"] Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.112835 4677 scope.go:117] "RemoveContainer" containerID="647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.198758 4677 scope.go:117] "RemoveContainer" containerID="91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075" Dec 03 15:28:48 crc kubenswrapper[4677]: E1203 15:28:48.199961 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075\": container with ID starting with 91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075 not found: ID does not exist" containerID="91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.200016 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075"} err="failed to get container status \"91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075\": rpc error: code = NotFound desc = could not find container \"91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075\": container with ID starting with 91eab9a63ede485258b1dcf7766d3cb98a71a978a2c86708ab262a68d611c075 not found: ID does not exist" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.200047 4677 scope.go:117] "RemoveContainer" containerID="6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116" Dec 03 15:28:48 crc kubenswrapper[4677]: E1203 15:28:48.200867 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116\": container with ID starting with 6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116 not found: ID does not exist" containerID="6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.200933 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116"} err="failed to get container status \"6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116\": rpc error: code = NotFound desc = could not find container \"6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116\": container with ID starting with 6f334e7bbc6a259c009bbf5739a8bfd9fddbc9c23bd9d7eda7b9a133bc1ae116 not found: ID does not exist" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.200989 4677 scope.go:117] "RemoveContainer" containerID="647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc" Dec 03 15:28:48 crc kubenswrapper[4677]: E1203 15:28:48.201329 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc\": container with ID starting with 647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc not found: ID does not exist" containerID="647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc" Dec 03 15:28:48 crc kubenswrapper[4677]: I1203 15:28:48.201370 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc"} err="failed to get container status \"647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc\": rpc error: code = NotFound desc = could not find container \"647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc\": container with ID starting with 647c169c23e63de3be0e2246ff2e2510511d93463c115cda46dc6c37da6e45fc not found: ID does not exist" Dec 03 15:28:49 crc kubenswrapper[4677]: I1203 15:28:49.995791 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" path="/var/lib/kubelet/pods/3062d8a6-830f-4daa-bc22-9381e8d00808/volumes" Dec 03 15:28:51 crc kubenswrapper[4677]: I1203 15:28:51.977096 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:28:51 crc kubenswrapper[4677]: E1203 15:28:51.977992 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:29:03 crc kubenswrapper[4677]: I1203 15:29:03.976648 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:29:03 crc kubenswrapper[4677]: E1203 15:29:03.977575 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:29:15 crc kubenswrapper[4677]: I1203 15:29:15.976731 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:29:15 crc kubenswrapper[4677]: E1203 15:29:15.977761 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:29:28 crc kubenswrapper[4677]: I1203 15:29:28.976646 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:29:28 crc kubenswrapper[4677]: E1203 15:29:28.979143 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:29:40 crc kubenswrapper[4677]: I1203 15:29:40.975817 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:29:41 crc kubenswrapper[4677]: I1203 15:29:41.629772 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"d4aee5cd748b1c99186ca94945a4630d59636700cf15498bdb5b82cf46e45b58"} Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.149369 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76"] Dec 03 15:30:00 crc kubenswrapper[4677]: E1203 15:30:00.150641 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.150660 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4677]: E1203 15:30:00.150716 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="extract-content" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.150725 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="extract-content" Dec 03 15:30:00 crc kubenswrapper[4677]: E1203 15:30:00.150752 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="extract-content" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.150784 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="extract-content" Dec 03 15:30:00 crc kubenswrapper[4677]: E1203 15:30:00.150802 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="extract-utilities" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.150810 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="extract-utilities" Dec 03 15:30:00 crc kubenswrapper[4677]: E1203 15:30:00.150825 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="extract-utilities" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.150833 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="extract-utilities" Dec 03 15:30:00 crc kubenswrapper[4677]: E1203 15:30:00.150872 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.150880 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.151227 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="3062d8a6-830f-4daa-bc22-9381e8d00808" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.151289 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="351fe843-d602-46b5-81ca-2468430d5eb1" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.152646 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.155401 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.155489 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.174094 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76"] Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.279655 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c799527-2b0c-4579-83f2-cf255f01550e-config-volume\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.279706 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c799527-2b0c-4579-83f2-cf255f01550e-secret-volume\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.279857 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fzmx\" (UniqueName: \"kubernetes.io/projected/7c799527-2b0c-4579-83f2-cf255f01550e-kube-api-access-4fzmx\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.382146 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c799527-2b0c-4579-83f2-cf255f01550e-config-volume\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.382229 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c799527-2b0c-4579-83f2-cf255f01550e-secret-volume\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.382395 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fzmx\" (UniqueName: \"kubernetes.io/projected/7c799527-2b0c-4579-83f2-cf255f01550e-kube-api-access-4fzmx\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.385635 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c799527-2b0c-4579-83f2-cf255f01550e-config-volume\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.391017 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c799527-2b0c-4579-83f2-cf255f01550e-secret-volume\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.402320 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fzmx\" (UniqueName: \"kubernetes.io/projected/7c799527-2b0c-4579-83f2-cf255f01550e-kube-api-access-4fzmx\") pod \"collect-profiles-29412930-brr76\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.474719 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:00 crc kubenswrapper[4677]: W1203 15:30:00.937448 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c799527_2b0c_4579_83f2_cf255f01550e.slice/crio-988c7b4f8d95035ce6627b89b909345dbcf92b2bddf151ad0f811d015b1f119d WatchSource:0}: Error finding container 988c7b4f8d95035ce6627b89b909345dbcf92b2bddf151ad0f811d015b1f119d: Status 404 returned error can't find the container with id 988c7b4f8d95035ce6627b89b909345dbcf92b2bddf151ad0f811d015b1f119d Dec 03 15:30:00 crc kubenswrapper[4677]: I1203 15:30:00.944629 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76"] Dec 03 15:30:01 crc kubenswrapper[4677]: I1203 15:30:01.861408 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" event={"ID":"7c799527-2b0c-4579-83f2-cf255f01550e","Type":"ContainerStarted","Data":"2f4c2a5fb6826f853114512afdc81240eade3c336001b2b9b0ab5808942d9dfe"} Dec 03 15:30:01 crc kubenswrapper[4677]: I1203 15:30:01.861725 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" event={"ID":"7c799527-2b0c-4579-83f2-cf255f01550e","Type":"ContainerStarted","Data":"988c7b4f8d95035ce6627b89b909345dbcf92b2bddf151ad0f811d015b1f119d"} Dec 03 15:30:01 crc kubenswrapper[4677]: I1203 15:30:01.893421 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" podStartSLOduration=1.893398575 podStartE2EDuration="1.893398575s" podCreationTimestamp="2025-12-03 15:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:30:01.884721318 +0000 UTC m=+6192.631053813" watchObservedRunningTime="2025-12-03 15:30:01.893398575 +0000 UTC m=+6192.639731030" Dec 03 15:30:02 crc kubenswrapper[4677]: I1203 15:30:02.876166 4677 generic.go:334] "Generic (PLEG): container finished" podID="7c799527-2b0c-4579-83f2-cf255f01550e" containerID="2f4c2a5fb6826f853114512afdc81240eade3c336001b2b9b0ab5808942d9dfe" exitCode=0 Dec 03 15:30:02 crc kubenswrapper[4677]: I1203 15:30:02.876224 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" event={"ID":"7c799527-2b0c-4579-83f2-cf255f01550e","Type":"ContainerDied","Data":"2f4c2a5fb6826f853114512afdc81240eade3c336001b2b9b0ab5808942d9dfe"} Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.276469 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.373252 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fzmx\" (UniqueName: \"kubernetes.io/projected/7c799527-2b0c-4579-83f2-cf255f01550e-kube-api-access-4fzmx\") pod \"7c799527-2b0c-4579-83f2-cf255f01550e\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.373347 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c799527-2b0c-4579-83f2-cf255f01550e-secret-volume\") pod \"7c799527-2b0c-4579-83f2-cf255f01550e\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.373372 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c799527-2b0c-4579-83f2-cf255f01550e-config-volume\") pod \"7c799527-2b0c-4579-83f2-cf255f01550e\" (UID: \"7c799527-2b0c-4579-83f2-cf255f01550e\") " Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.374142 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c799527-2b0c-4579-83f2-cf255f01550e-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c799527-2b0c-4579-83f2-cf255f01550e" (UID: "7c799527-2b0c-4579-83f2-cf255f01550e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.381679 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c799527-2b0c-4579-83f2-cf255f01550e-kube-api-access-4fzmx" (OuterVolumeSpecName: "kube-api-access-4fzmx") pod "7c799527-2b0c-4579-83f2-cf255f01550e" (UID: "7c799527-2b0c-4579-83f2-cf255f01550e"). InnerVolumeSpecName "kube-api-access-4fzmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.391322 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c799527-2b0c-4579-83f2-cf255f01550e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c799527-2b0c-4579-83f2-cf255f01550e" (UID: "7c799527-2b0c-4579-83f2-cf255f01550e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.475411 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c799527-2b0c-4579-83f2-cf255f01550e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.475464 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c799527-2b0c-4579-83f2-cf255f01550e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.475486 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fzmx\" (UniqueName: \"kubernetes.io/projected/7c799527-2b0c-4579-83f2-cf255f01550e-kube-api-access-4fzmx\") on node \"crc\" DevicePath \"\"" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.933698 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" event={"ID":"7c799527-2b0c-4579-83f2-cf255f01550e","Type":"ContainerDied","Data":"988c7b4f8d95035ce6627b89b909345dbcf92b2bddf151ad0f811d015b1f119d"} Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.933762 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="988c7b4f8d95035ce6627b89b909345dbcf92b2bddf151ad0f811d015b1f119d" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.933773 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-brr76" Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.980038 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds"] Dec 03 15:30:04 crc kubenswrapper[4677]: I1203 15:30:04.988156 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-5x9ds"] Dec 03 15:30:05 crc kubenswrapper[4677]: I1203 15:30:05.987439 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8213e335-c755-42a3-aa53-f901c7cc24b1" path="/var/lib/kubelet/pods/8213e335-c755-42a3-aa53-f901c7cc24b1/volumes" Dec 03 15:30:52 crc kubenswrapper[4677]: I1203 15:30:52.145496 4677 scope.go:117] "RemoveContainer" containerID="6f3585ef9ef7b61ec0bbb1e6b567f9720ddfe2552aa45c5a90419e9fd5657dce" Dec 03 15:32:08 crc kubenswrapper[4677]: I1203 15:32:08.436965 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:32:08 crc kubenswrapper[4677]: I1203 15:32:08.437547 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:32:38 crc kubenswrapper[4677]: I1203 15:32:38.437813 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:32:38 crc kubenswrapper[4677]: I1203 15:32:38.438695 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.437987 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.438639 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.438722 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.439737 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4aee5cd748b1c99186ca94945a4630d59636700cf15498bdb5b82cf46e45b58"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.439841 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://d4aee5cd748b1c99186ca94945a4630d59636700cf15498bdb5b82cf46e45b58" gracePeriod=600 Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.951390 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="d4aee5cd748b1c99186ca94945a4630d59636700cf15498bdb5b82cf46e45b58" exitCode=0 Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.951423 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"d4aee5cd748b1c99186ca94945a4630d59636700cf15498bdb5b82cf46e45b58"} Dec 03 15:33:08 crc kubenswrapper[4677]: I1203 15:33:08.951845 4677 scope.go:117] "RemoveContainer" containerID="c6c9dd4b8f02aabbbdc85eeb199b0755a2ec0430d042cd0fabe78ffeb5207c9b" Dec 03 15:33:09 crc kubenswrapper[4677]: I1203 15:33:09.963942 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27"} Dec 03 15:35:38 crc kubenswrapper[4677]: I1203 15:35:38.437778 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:35:38 crc kubenswrapper[4677]: I1203 15:35:38.438661 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:36:08 crc kubenswrapper[4677]: I1203 15:36:08.437384 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:36:08 crc kubenswrapper[4677]: I1203 15:36:08.439268 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:36:38 crc kubenswrapper[4677]: I1203 15:36:38.436875 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:36:38 crc kubenswrapper[4677]: I1203 15:36:38.437470 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:36:38 crc kubenswrapper[4677]: I1203 15:36:38.437521 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:36:38 crc kubenswrapper[4677]: I1203 15:36:38.438433 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:36:38 crc kubenswrapper[4677]: I1203 15:36:38.438501 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" gracePeriod=600 Dec 03 15:36:38 crc kubenswrapper[4677]: E1203 15:36:38.570155 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.269477 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t7xsl"] Dec 03 15:36:39 crc kubenswrapper[4677]: E1203 15:36:39.270641 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c799527-2b0c-4579-83f2-cf255f01550e" containerName="collect-profiles" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.270674 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c799527-2b0c-4579-83f2-cf255f01550e" containerName="collect-profiles" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.271130 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c799527-2b0c-4579-83f2-cf255f01550e" containerName="collect-profiles" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.272697 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" exitCode=0 Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.273353 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27"} Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.273411 4677 scope.go:117] "RemoveContainer" containerID="d4aee5cd748b1c99186ca94945a4630d59636700cf15498bdb5b82cf46e45b58" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.274198 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:36:39 crc kubenswrapper[4677]: E1203 15:36:39.274506 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.274784 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.283896 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7xsl"] Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.339434 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-utilities\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.339518 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prd8g\" (UniqueName: \"kubernetes.io/projected/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-kube-api-access-prd8g\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.339543 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-catalog-content\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.441508 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prd8g\" (UniqueName: \"kubernetes.io/projected/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-kube-api-access-prd8g\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.441578 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-catalog-content\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.441823 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-utilities\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.442178 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-catalog-content\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.442291 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-utilities\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.473683 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prd8g\" (UniqueName: \"kubernetes.io/projected/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-kube-api-access-prd8g\") pod \"redhat-operators-t7xsl\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:39 crc kubenswrapper[4677]: I1203 15:36:39.620678 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:40 crc kubenswrapper[4677]: I1203 15:36:40.152100 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7xsl"] Dec 03 15:36:40 crc kubenswrapper[4677]: I1203 15:36:40.286332 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7xsl" event={"ID":"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3","Type":"ContainerStarted","Data":"2067f2cefb4ad81fbd318709114419a7fae8160d70d9037eea51d41f2aa61815"} Dec 03 15:36:41 crc kubenswrapper[4677]: I1203 15:36:41.302651 4677 generic.go:334] "Generic (PLEG): container finished" podID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerID="6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb" exitCode=0 Dec 03 15:36:41 crc kubenswrapper[4677]: I1203 15:36:41.302834 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7xsl" event={"ID":"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3","Type":"ContainerDied","Data":"6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb"} Dec 03 15:36:41 crc kubenswrapper[4677]: I1203 15:36:41.306081 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:36:42 crc kubenswrapper[4677]: I1203 15:36:42.318244 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7xsl" event={"ID":"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3","Type":"ContainerStarted","Data":"5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a"} Dec 03 15:36:45 crc kubenswrapper[4677]: I1203 15:36:45.350096 4677 generic.go:334] "Generic (PLEG): container finished" podID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerID="5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a" exitCode=0 Dec 03 15:36:45 crc kubenswrapper[4677]: I1203 15:36:45.350229 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7xsl" event={"ID":"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3","Type":"ContainerDied","Data":"5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a"} Dec 03 15:36:46 crc kubenswrapper[4677]: I1203 15:36:46.363651 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7xsl" event={"ID":"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3","Type":"ContainerStarted","Data":"c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108"} Dec 03 15:36:46 crc kubenswrapper[4677]: I1203 15:36:46.381088 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t7xsl" podStartSLOduration=2.872715938 podStartE2EDuration="7.381063714s" podCreationTimestamp="2025-12-03 15:36:39 +0000 UTC" firstStartedPulling="2025-12-03 15:36:41.305722665 +0000 UTC m=+6592.052055140" lastFinishedPulling="2025-12-03 15:36:45.814070421 +0000 UTC m=+6596.560402916" observedRunningTime="2025-12-03 15:36:46.378742451 +0000 UTC m=+6597.125074936" watchObservedRunningTime="2025-12-03 15:36:46.381063714 +0000 UTC m=+6597.127396169" Dec 03 15:36:49 crc kubenswrapper[4677]: I1203 15:36:49.622381 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:49 crc kubenswrapper[4677]: I1203 15:36:49.622805 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:50 crc kubenswrapper[4677]: I1203 15:36:50.686573 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t7xsl" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="registry-server" probeResult="failure" output=< Dec 03 15:36:50 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 15:36:50 crc kubenswrapper[4677]: > Dec 03 15:36:52 crc kubenswrapper[4677]: I1203 15:36:52.976210 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:36:52 crc kubenswrapper[4677]: E1203 15:36:52.977857 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:36:59 crc kubenswrapper[4677]: I1203 15:36:59.670097 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:59 crc kubenswrapper[4677]: I1203 15:36:59.735443 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:36:59 crc kubenswrapper[4677]: I1203 15:36:59.914635 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7xsl"] Dec 03 15:37:01 crc kubenswrapper[4677]: I1203 15:37:01.542858 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t7xsl" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="registry-server" containerID="cri-o://c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108" gracePeriod=2 Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.074367 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.245241 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-catalog-content\") pod \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.245598 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prd8g\" (UniqueName: \"kubernetes.io/projected/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-kube-api-access-prd8g\") pod \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.245641 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-utilities\") pod \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\" (UID: \"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3\") " Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.246589 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-utilities" (OuterVolumeSpecName: "utilities") pod "a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" (UID: "a88704bb-65b8-4ec2-8992-44e6d2b9cbe3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.257716 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-kube-api-access-prd8g" (OuterVolumeSpecName: "kube-api-access-prd8g") pod "a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" (UID: "a88704bb-65b8-4ec2-8992-44e6d2b9cbe3"). InnerVolumeSpecName "kube-api-access-prd8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.347886 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prd8g\" (UniqueName: \"kubernetes.io/projected/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-kube-api-access-prd8g\") on node \"crc\" DevicePath \"\"" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.347919 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.382693 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" (UID: "a88704bb-65b8-4ec2-8992-44e6d2b9cbe3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.449649 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.555191 4677 generic.go:334] "Generic (PLEG): container finished" podID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerID="c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108" exitCode=0 Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.555234 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7xsl" event={"ID":"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3","Type":"ContainerDied","Data":"c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108"} Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.555248 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7xsl" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.555275 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7xsl" event={"ID":"a88704bb-65b8-4ec2-8992-44e6d2b9cbe3","Type":"ContainerDied","Data":"2067f2cefb4ad81fbd318709114419a7fae8160d70d9037eea51d41f2aa61815"} Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.555299 4677 scope.go:117] "RemoveContainer" containerID="c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.588830 4677 scope.go:117] "RemoveContainer" containerID="5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.591510 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7xsl"] Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.603886 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t7xsl"] Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.609044 4677 scope.go:117] "RemoveContainer" containerID="6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.670827 4677 scope.go:117] "RemoveContainer" containerID="c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108" Dec 03 15:37:02 crc kubenswrapper[4677]: E1203 15:37:02.671296 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108\": container with ID starting with c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108 not found: ID does not exist" containerID="c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.671353 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108"} err="failed to get container status \"c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108\": rpc error: code = NotFound desc = could not find container \"c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108\": container with ID starting with c548612b28834018f869629710fb3c7b0023b34bf9072ba8bb89626ba44bc108 not found: ID does not exist" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.671381 4677 scope.go:117] "RemoveContainer" containerID="5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a" Dec 03 15:37:02 crc kubenswrapper[4677]: E1203 15:37:02.671616 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a\": container with ID starting with 5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a not found: ID does not exist" containerID="5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.671637 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a"} err="failed to get container status \"5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a\": rpc error: code = NotFound desc = could not find container \"5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a\": container with ID starting with 5414dc5229cbdbe30e93988e6607848c76e3a81c50b8b9fa9ed0eea37ecdd79a not found: ID does not exist" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.671650 4677 scope.go:117] "RemoveContainer" containerID="6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb" Dec 03 15:37:02 crc kubenswrapper[4677]: E1203 15:37:02.671873 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb\": container with ID starting with 6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb not found: ID does not exist" containerID="6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb" Dec 03 15:37:02 crc kubenswrapper[4677]: I1203 15:37:02.671894 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb"} err="failed to get container status \"6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb\": rpc error: code = NotFound desc = could not find container \"6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb\": container with ID starting with 6518445e4cf2f0937a8f0873932c240e9f3a5b6da549b90c5a55d8a0fb47a1bb not found: ID does not exist" Dec 03 15:37:03 crc kubenswrapper[4677]: I1203 15:37:03.992159 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" path="/var/lib/kubelet/pods/a88704bb-65b8-4ec2-8992-44e6d2b9cbe3/volumes" Dec 03 15:37:06 crc kubenswrapper[4677]: I1203 15:37:06.976768 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:37:06 crc kubenswrapper[4677]: E1203 15:37:06.978023 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:37:20 crc kubenswrapper[4677]: I1203 15:37:20.976473 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:37:20 crc kubenswrapper[4677]: E1203 15:37:20.977344 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:37:35 crc kubenswrapper[4677]: I1203 15:37:35.976494 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:37:35 crc kubenswrapper[4677]: E1203 15:37:35.977419 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:37:49 crc kubenswrapper[4677]: I1203 15:37:49.982832 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:37:49 crc kubenswrapper[4677]: E1203 15:37:49.983755 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:38:01 crc kubenswrapper[4677]: I1203 15:38:01.976812 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:38:01 crc kubenswrapper[4677]: E1203 15:38:01.977801 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:38:16 crc kubenswrapper[4677]: I1203 15:38:16.127640 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:38:16 crc kubenswrapper[4677]: E1203 15:38:16.128490 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:38:28 crc kubenswrapper[4677]: I1203 15:38:28.976808 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:38:28 crc kubenswrapper[4677]: E1203 15:38:28.977705 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.287678 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f9k2j"] Dec 03 15:38:41 crc kubenswrapper[4677]: E1203 15:38:41.289560 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="extract-utilities" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.289607 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="extract-utilities" Dec 03 15:38:41 crc kubenswrapper[4677]: E1203 15:38:41.289648 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="extract-content" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.289667 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="extract-content" Dec 03 15:38:41 crc kubenswrapper[4677]: E1203 15:38:41.289718 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="registry-server" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.289737 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="registry-server" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.290364 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a88704bb-65b8-4ec2-8992-44e6d2b9cbe3" containerName="registry-server" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.298414 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.310687 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9k2j"] Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.406154 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-utilities\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.406716 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-catalog-content\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.407446 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcxlc\" (UniqueName: \"kubernetes.io/projected/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-kube-api-access-lcxlc\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.510374 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-catalog-content\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.510473 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcxlc\" (UniqueName: \"kubernetes.io/projected/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-kube-api-access-lcxlc\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.510608 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-utilities\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.511029 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-utilities\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.511072 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-catalog-content\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.535162 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcxlc\" (UniqueName: \"kubernetes.io/projected/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-kube-api-access-lcxlc\") pod \"redhat-marketplace-f9k2j\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:41 crc kubenswrapper[4677]: I1203 15:38:41.633767 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:42 crc kubenswrapper[4677]: I1203 15:38:42.140871 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9k2j"] Dec 03 15:38:42 crc kubenswrapper[4677]: I1203 15:38:42.688679 4677 generic.go:334] "Generic (PLEG): container finished" podID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerID="1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243" exitCode=0 Dec 03 15:38:42 crc kubenswrapper[4677]: I1203 15:38:42.688758 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9k2j" event={"ID":"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7","Type":"ContainerDied","Data":"1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243"} Dec 03 15:38:42 crc kubenswrapper[4677]: I1203 15:38:42.689056 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9k2j" event={"ID":"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7","Type":"ContainerStarted","Data":"007cb233a6297bc8044a8ed49a5a2d1c768aa3d29522439e18f5d0c3016a31ea"} Dec 03 15:38:42 crc kubenswrapper[4677]: I1203 15:38:42.977198 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:38:42 crc kubenswrapper[4677]: E1203 15:38:42.977887 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:38:43 crc kubenswrapper[4677]: I1203 15:38:43.706380 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9k2j" event={"ID":"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7","Type":"ContainerStarted","Data":"4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c"} Dec 03 15:38:44 crc kubenswrapper[4677]: I1203 15:38:44.720052 4677 generic.go:334] "Generic (PLEG): container finished" podID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerID="4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c" exitCode=0 Dec 03 15:38:44 crc kubenswrapper[4677]: I1203 15:38:44.720130 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9k2j" event={"ID":"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7","Type":"ContainerDied","Data":"4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c"} Dec 03 15:38:45 crc kubenswrapper[4677]: I1203 15:38:45.731698 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9k2j" event={"ID":"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7","Type":"ContainerStarted","Data":"270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281"} Dec 03 15:38:45 crc kubenswrapper[4677]: I1203 15:38:45.764839 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f9k2j" podStartSLOduration=2.252895351 podStartE2EDuration="4.764818555s" podCreationTimestamp="2025-12-03 15:38:41 +0000 UTC" firstStartedPulling="2025-12-03 15:38:42.690857862 +0000 UTC m=+6713.437190317" lastFinishedPulling="2025-12-03 15:38:45.202781066 +0000 UTC m=+6715.949113521" observedRunningTime="2025-12-03 15:38:45.758574997 +0000 UTC m=+6716.504907452" watchObservedRunningTime="2025-12-03 15:38:45.764818555 +0000 UTC m=+6716.511151030" Dec 03 15:38:51 crc kubenswrapper[4677]: I1203 15:38:51.634642 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:51 crc kubenswrapper[4677]: I1203 15:38:51.636705 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:51 crc kubenswrapper[4677]: I1203 15:38:51.697715 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:51 crc kubenswrapper[4677]: I1203 15:38:51.865300 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:51 crc kubenswrapper[4677]: I1203 15:38:51.946187 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9k2j"] Dec 03 15:38:53 crc kubenswrapper[4677]: I1203 15:38:53.821648 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f9k2j" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="registry-server" containerID="cri-o://270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281" gracePeriod=2 Dec 03 15:38:53 crc kubenswrapper[4677]: I1203 15:38:53.978057 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:38:53 crc kubenswrapper[4677]: E1203 15:38:53.978720 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.386731 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.560368 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcxlc\" (UniqueName: \"kubernetes.io/projected/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-kube-api-access-lcxlc\") pod \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.560448 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-catalog-content\") pod \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.560593 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-utilities\") pod \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\" (UID: \"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7\") " Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.561860 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-utilities" (OuterVolumeSpecName: "utilities") pod "f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" (UID: "f0581fef-dbf0-4ffd-bc9b-b08954abb7a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.568108 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-kube-api-access-lcxlc" (OuterVolumeSpecName: "kube-api-access-lcxlc") pod "f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" (UID: "f0581fef-dbf0-4ffd-bc9b-b08954abb7a7"). InnerVolumeSpecName "kube-api-access-lcxlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.582743 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" (UID: "f0581fef-dbf0-4ffd-bc9b-b08954abb7a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.663200 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcxlc\" (UniqueName: \"kubernetes.io/projected/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-kube-api-access-lcxlc\") on node \"crc\" DevicePath \"\"" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.663250 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.663266 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.835064 4677 generic.go:334] "Generic (PLEG): container finished" podID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerID="270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281" exitCode=0 Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.835138 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9k2j" event={"ID":"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7","Type":"ContainerDied","Data":"270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281"} Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.835237 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9k2j" event={"ID":"f0581fef-dbf0-4ffd-bc9b-b08954abb7a7","Type":"ContainerDied","Data":"007cb233a6297bc8044a8ed49a5a2d1c768aa3d29522439e18f5d0c3016a31ea"} Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.835275 4677 scope.go:117] "RemoveContainer" containerID="270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.835164 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9k2j" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.882902 4677 scope.go:117] "RemoveContainer" containerID="4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.892448 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9k2j"] Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.903548 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9k2j"] Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.909764 4677 scope.go:117] "RemoveContainer" containerID="1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.960593 4677 scope.go:117] "RemoveContainer" containerID="270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281" Dec 03 15:38:54 crc kubenswrapper[4677]: E1203 15:38:54.961197 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281\": container with ID starting with 270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281 not found: ID does not exist" containerID="270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.961244 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281"} err="failed to get container status \"270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281\": rpc error: code = NotFound desc = could not find container \"270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281\": container with ID starting with 270254f31708e0ee6a3200f0b1b26bf6bebde5bbaaeaf07b7a4fd9c7acefa281 not found: ID does not exist" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.961273 4677 scope.go:117] "RemoveContainer" containerID="4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c" Dec 03 15:38:54 crc kubenswrapper[4677]: E1203 15:38:54.961852 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c\": container with ID starting with 4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c not found: ID does not exist" containerID="4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.961914 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c"} err="failed to get container status \"4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c\": rpc error: code = NotFound desc = could not find container \"4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c\": container with ID starting with 4935714740fe35a3ad428816b8e0cad4827ae1a8c7575e7d934fb664c147a00c not found: ID does not exist" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.962004 4677 scope.go:117] "RemoveContainer" containerID="1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243" Dec 03 15:38:54 crc kubenswrapper[4677]: E1203 15:38:54.962530 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243\": container with ID starting with 1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243 not found: ID does not exist" containerID="1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243" Dec 03 15:38:54 crc kubenswrapper[4677]: I1203 15:38:54.962561 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243"} err="failed to get container status \"1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243\": rpc error: code = NotFound desc = could not find container \"1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243\": container with ID starting with 1920cfd27997c5d37e7085cfd09fae6fc665836577e739f1a5a0bff30961a243 not found: ID does not exist" Dec 03 15:38:55 crc kubenswrapper[4677]: I1203 15:38:55.997823 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" path="/var/lib/kubelet/pods/f0581fef-dbf0-4ffd-bc9b-b08954abb7a7/volumes" Dec 03 15:39:08 crc kubenswrapper[4677]: I1203 15:39:08.976439 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:39:08 crc kubenswrapper[4677]: E1203 15:39:08.977424 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:39:23 crc kubenswrapper[4677]: I1203 15:39:23.976169 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:39:23 crc kubenswrapper[4677]: E1203 15:39:23.977066 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:39:35 crc kubenswrapper[4677]: I1203 15:39:35.977005 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:39:35 crc kubenswrapper[4677]: E1203 15:39:35.977906 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:39:50 crc kubenswrapper[4677]: I1203 15:39:50.976017 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:39:50 crc kubenswrapper[4677]: E1203 15:39:50.976717 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:40:03 crc kubenswrapper[4677]: I1203 15:40:03.976416 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:40:03 crc kubenswrapper[4677]: E1203 15:40:03.977295 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.654108 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rsqkb"] Dec 03 15:40:11 crc kubenswrapper[4677]: E1203 15:40:11.655508 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="registry-server" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.655527 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="registry-server" Dec 03 15:40:11 crc kubenswrapper[4677]: E1203 15:40:11.655560 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="extract-content" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.655568 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="extract-content" Dec 03 15:40:11 crc kubenswrapper[4677]: E1203 15:40:11.655582 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="extract-utilities" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.655590 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="extract-utilities" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.655800 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0581fef-dbf0-4ffd-bc9b-b08954abb7a7" containerName="registry-server" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.657596 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.678752 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rsqkb"] Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.703241 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-catalog-content\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.703455 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jfpr\" (UniqueName: \"kubernetes.io/projected/88150969-a1fa-439e-9719-ba331eb2f657-kube-api-access-6jfpr\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.703563 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-utilities\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.805514 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jfpr\" (UniqueName: \"kubernetes.io/projected/88150969-a1fa-439e-9719-ba331eb2f657-kube-api-access-6jfpr\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.805602 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-utilities\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.805747 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-catalog-content\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.806144 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-utilities\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.806237 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-catalog-content\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.847709 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jfpr\" (UniqueName: \"kubernetes.io/projected/88150969-a1fa-439e-9719-ba331eb2f657-kube-api-access-6jfpr\") pod \"certified-operators-rsqkb\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:11 crc kubenswrapper[4677]: I1203 15:40:11.978701 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:12 crc kubenswrapper[4677]: I1203 15:40:12.590467 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rsqkb"] Dec 03 15:40:12 crc kubenswrapper[4677]: I1203 15:40:12.720064 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsqkb" event={"ID":"88150969-a1fa-439e-9719-ba331eb2f657","Type":"ContainerStarted","Data":"51448d1f22dc0040bafdd100e2ed1f8fd557e37ae9a86b2716e2615145510023"} Dec 03 15:40:13 crc kubenswrapper[4677]: I1203 15:40:13.734592 4677 generic.go:334] "Generic (PLEG): container finished" podID="88150969-a1fa-439e-9719-ba331eb2f657" containerID="59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30" exitCode=0 Dec 03 15:40:13 crc kubenswrapper[4677]: I1203 15:40:13.734693 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsqkb" event={"ID":"88150969-a1fa-439e-9719-ba331eb2f657","Type":"ContainerDied","Data":"59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30"} Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.066003 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jwhmf"] Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.071429 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.083209 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwhmf"] Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.176251 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gfq2\" (UniqueName: \"kubernetes.io/projected/a16d95e4-0a07-4c8e-a023-561224f1c78c-kube-api-access-9gfq2\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.176306 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-utilities\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.176423 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-catalog-content\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.278774 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-catalog-content\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.278892 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gfq2\" (UniqueName: \"kubernetes.io/projected/a16d95e4-0a07-4c8e-a023-561224f1c78c-kube-api-access-9gfq2\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.278921 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-utilities\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.279445 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-utilities\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.279700 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-catalog-content\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.306217 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gfq2\" (UniqueName: \"kubernetes.io/projected/a16d95e4-0a07-4c8e-a023-561224f1c78c-kube-api-access-9gfq2\") pod \"community-operators-jwhmf\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.402384 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:14 crc kubenswrapper[4677]: I1203 15:40:14.976393 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:40:14 crc kubenswrapper[4677]: E1203 15:40:14.976673 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:40:15 crc kubenswrapper[4677]: I1203 15:40:15.083186 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwhmf"] Dec 03 15:40:15 crc kubenswrapper[4677]: I1203 15:40:15.755584 4677 generic.go:334] "Generic (PLEG): container finished" podID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerID="bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013" exitCode=0 Dec 03 15:40:15 crc kubenswrapper[4677]: I1203 15:40:15.755628 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwhmf" event={"ID":"a16d95e4-0a07-4c8e-a023-561224f1c78c","Type":"ContainerDied","Data":"bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013"} Dec 03 15:40:15 crc kubenswrapper[4677]: I1203 15:40:15.756025 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwhmf" event={"ID":"a16d95e4-0a07-4c8e-a023-561224f1c78c","Type":"ContainerStarted","Data":"32bc9a9c0ef63a1464c7bb65db3848f7fa683795aa72291e32156a3182797980"} Dec 03 15:40:15 crc kubenswrapper[4677]: I1203 15:40:15.758134 4677 generic.go:334] "Generic (PLEG): container finished" podID="88150969-a1fa-439e-9719-ba331eb2f657" containerID="eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0" exitCode=0 Dec 03 15:40:15 crc kubenswrapper[4677]: I1203 15:40:15.758166 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsqkb" event={"ID":"88150969-a1fa-439e-9719-ba331eb2f657","Type":"ContainerDied","Data":"eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0"} Dec 03 15:40:16 crc kubenswrapper[4677]: I1203 15:40:16.771137 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwhmf" event={"ID":"a16d95e4-0a07-4c8e-a023-561224f1c78c","Type":"ContainerStarted","Data":"bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a"} Dec 03 15:40:16 crc kubenswrapper[4677]: I1203 15:40:16.776355 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsqkb" event={"ID":"88150969-a1fa-439e-9719-ba331eb2f657","Type":"ContainerStarted","Data":"5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be"} Dec 03 15:40:16 crc kubenswrapper[4677]: I1203 15:40:16.824368 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rsqkb" podStartSLOduration=3.435727656 podStartE2EDuration="5.824349668s" podCreationTimestamp="2025-12-03 15:40:11 +0000 UTC" firstStartedPulling="2025-12-03 15:40:13.737040764 +0000 UTC m=+6804.483373229" lastFinishedPulling="2025-12-03 15:40:16.125662786 +0000 UTC m=+6806.871995241" observedRunningTime="2025-12-03 15:40:16.818242923 +0000 UTC m=+6807.564575388" watchObservedRunningTime="2025-12-03 15:40:16.824349668 +0000 UTC m=+6807.570682123" Dec 03 15:40:17 crc kubenswrapper[4677]: I1203 15:40:17.787828 4677 generic.go:334] "Generic (PLEG): container finished" podID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerID="bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a" exitCode=0 Dec 03 15:40:17 crc kubenswrapper[4677]: I1203 15:40:17.787923 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwhmf" event={"ID":"a16d95e4-0a07-4c8e-a023-561224f1c78c","Type":"ContainerDied","Data":"bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a"} Dec 03 15:40:18 crc kubenswrapper[4677]: I1203 15:40:18.804310 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwhmf" event={"ID":"a16d95e4-0a07-4c8e-a023-561224f1c78c","Type":"ContainerStarted","Data":"48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9"} Dec 03 15:40:18 crc kubenswrapper[4677]: I1203 15:40:18.828136 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jwhmf" podStartSLOduration=2.413646998 podStartE2EDuration="4.828119079s" podCreationTimestamp="2025-12-03 15:40:14 +0000 UTC" firstStartedPulling="2025-12-03 15:40:15.758802082 +0000 UTC m=+6806.505134537" lastFinishedPulling="2025-12-03 15:40:18.173274123 +0000 UTC m=+6808.919606618" observedRunningTime="2025-12-03 15:40:18.819041724 +0000 UTC m=+6809.565374249" watchObservedRunningTime="2025-12-03 15:40:18.828119079 +0000 UTC m=+6809.574451534" Dec 03 15:40:21 crc kubenswrapper[4677]: I1203 15:40:21.988032 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:21 crc kubenswrapper[4677]: I1203 15:40:21.988086 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:22 crc kubenswrapper[4677]: I1203 15:40:22.028201 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:22 crc kubenswrapper[4677]: I1203 15:40:22.911660 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:23 crc kubenswrapper[4677]: I1203 15:40:23.449071 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rsqkb"] Dec 03 15:40:24 crc kubenswrapper[4677]: I1203 15:40:24.403173 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:24 crc kubenswrapper[4677]: I1203 15:40:24.403443 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:24 crc kubenswrapper[4677]: I1203 15:40:24.450905 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:24 crc kubenswrapper[4677]: I1203 15:40:24.890747 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rsqkb" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="registry-server" containerID="cri-o://5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be" gracePeriod=2 Dec 03 15:40:24 crc kubenswrapper[4677]: I1203 15:40:24.947235 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.384307 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.453524 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-catalog-content\") pod \"88150969-a1fa-439e-9719-ba331eb2f657\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.453746 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-utilities\") pod \"88150969-a1fa-439e-9719-ba331eb2f657\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.454076 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jfpr\" (UniqueName: \"kubernetes.io/projected/88150969-a1fa-439e-9719-ba331eb2f657-kube-api-access-6jfpr\") pod \"88150969-a1fa-439e-9719-ba331eb2f657\" (UID: \"88150969-a1fa-439e-9719-ba331eb2f657\") " Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.454743 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-utilities" (OuterVolumeSpecName: "utilities") pod "88150969-a1fa-439e-9719-ba331eb2f657" (UID: "88150969-a1fa-439e-9719-ba331eb2f657"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.462393 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88150969-a1fa-439e-9719-ba331eb2f657-kube-api-access-6jfpr" (OuterVolumeSpecName: "kube-api-access-6jfpr") pod "88150969-a1fa-439e-9719-ba331eb2f657" (UID: "88150969-a1fa-439e-9719-ba331eb2f657"). InnerVolumeSpecName "kube-api-access-6jfpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.502914 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88150969-a1fa-439e-9719-ba331eb2f657" (UID: "88150969-a1fa-439e-9719-ba331eb2f657"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.556584 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jfpr\" (UniqueName: \"kubernetes.io/projected/88150969-a1fa-439e-9719-ba331eb2f657-kube-api-access-6jfpr\") on node \"crc\" DevicePath \"\"" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.556815 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.556875 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88150969-a1fa-439e-9719-ba331eb2f657-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.906198 4677 generic.go:334] "Generic (PLEG): container finished" podID="88150969-a1fa-439e-9719-ba331eb2f657" containerID="5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be" exitCode=0 Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.906310 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsqkb" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.906370 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsqkb" event={"ID":"88150969-a1fa-439e-9719-ba331eb2f657","Type":"ContainerDied","Data":"5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be"} Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.906410 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsqkb" event={"ID":"88150969-a1fa-439e-9719-ba331eb2f657","Type":"ContainerDied","Data":"51448d1f22dc0040bafdd100e2ed1f8fd557e37ae9a86b2716e2615145510023"} Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.906431 4677 scope.go:117] "RemoveContainer" containerID="5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.934110 4677 scope.go:117] "RemoveContainer" containerID="eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.959558 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rsqkb"] Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.974471 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rsqkb"] Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.982773 4677 scope.go:117] "RemoveContainer" containerID="59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30" Dec 03 15:40:25 crc kubenswrapper[4677]: I1203 15:40:25.991847 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88150969-a1fa-439e-9719-ba331eb2f657" path="/var/lib/kubelet/pods/88150969-a1fa-439e-9719-ba331eb2f657/volumes" Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.043753 4677 scope.go:117] "RemoveContainer" containerID="5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be" Dec 03 15:40:26 crc kubenswrapper[4677]: E1203 15:40:26.045252 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be\": container with ID starting with 5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be not found: ID does not exist" containerID="5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be" Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.045431 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be"} err="failed to get container status \"5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be\": rpc error: code = NotFound desc = could not find container \"5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be\": container with ID starting with 5da5034131b94a08e9c3a253f414b1a7141827d51af4721576f7d2d8aac7f6be not found: ID does not exist" Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.045492 4677 scope.go:117] "RemoveContainer" containerID="eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0" Dec 03 15:40:26 crc kubenswrapper[4677]: E1203 15:40:26.049208 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0\": container with ID starting with eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0 not found: ID does not exist" containerID="eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0" Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.049260 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0"} err="failed to get container status \"eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0\": rpc error: code = NotFound desc = could not find container \"eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0\": container with ID starting with eef4a28dbb5230cd92ddb233f2af780096de36512fe2c4afc3037c046850ced0 not found: ID does not exist" Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.049295 4677 scope.go:117] "RemoveContainer" containerID="59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30" Dec 03 15:40:26 crc kubenswrapper[4677]: E1203 15:40:26.052317 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30\": container with ID starting with 59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30 not found: ID does not exist" containerID="59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30" Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.052362 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30"} err="failed to get container status \"59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30\": rpc error: code = NotFound desc = could not find container \"59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30\": container with ID starting with 59a467229a72e8c96a6e3528e563a71af88b931c598b75d324cbf212c542bc30 not found: ID does not exist" Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.252585 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jwhmf"] Dec 03 15:40:26 crc kubenswrapper[4677]: I1203 15:40:26.918966 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jwhmf" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="registry-server" containerID="cri-o://48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9" gracePeriod=2 Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.381805 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.499245 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gfq2\" (UniqueName: \"kubernetes.io/projected/a16d95e4-0a07-4c8e-a023-561224f1c78c-kube-api-access-9gfq2\") pod \"a16d95e4-0a07-4c8e-a023-561224f1c78c\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.499474 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-catalog-content\") pod \"a16d95e4-0a07-4c8e-a023-561224f1c78c\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.499498 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-utilities\") pod \"a16d95e4-0a07-4c8e-a023-561224f1c78c\" (UID: \"a16d95e4-0a07-4c8e-a023-561224f1c78c\") " Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.500531 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-utilities" (OuterVolumeSpecName: "utilities") pod "a16d95e4-0a07-4c8e-a023-561224f1c78c" (UID: "a16d95e4-0a07-4c8e-a023-561224f1c78c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.503931 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a16d95e4-0a07-4c8e-a023-561224f1c78c-kube-api-access-9gfq2" (OuterVolumeSpecName: "kube-api-access-9gfq2") pod "a16d95e4-0a07-4c8e-a023-561224f1c78c" (UID: "a16d95e4-0a07-4c8e-a023-561224f1c78c"). InnerVolumeSpecName "kube-api-access-9gfq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.568311 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a16d95e4-0a07-4c8e-a023-561224f1c78c" (UID: "a16d95e4-0a07-4c8e-a023-561224f1c78c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.602617 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.602663 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a16d95e4-0a07-4c8e-a023-561224f1c78c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.602676 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gfq2\" (UniqueName: \"kubernetes.io/projected/a16d95e4-0a07-4c8e-a023-561224f1c78c-kube-api-access-9gfq2\") on node \"crc\" DevicePath \"\"" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.936765 4677 generic.go:334] "Generic (PLEG): container finished" podID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerID="48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9" exitCode=0 Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.936867 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwhmf" event={"ID":"a16d95e4-0a07-4c8e-a023-561224f1c78c","Type":"ContainerDied","Data":"48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9"} Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.936944 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwhmf" event={"ID":"a16d95e4-0a07-4c8e-a023-561224f1c78c","Type":"ContainerDied","Data":"32bc9a9c0ef63a1464c7bb65db3848f7fa683795aa72291e32156a3182797980"} Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.936885 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwhmf" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.937005 4677 scope.go:117] "RemoveContainer" containerID="48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9" Dec 03 15:40:27 crc kubenswrapper[4677]: I1203 15:40:27.988551 4677 scope.go:117] "RemoveContainer" containerID="bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a" Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.007594 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jwhmf"] Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.007940 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jwhmf"] Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.032487 4677 scope.go:117] "RemoveContainer" containerID="bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013" Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.065809 4677 scope.go:117] "RemoveContainer" containerID="48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9" Dec 03 15:40:28 crc kubenswrapper[4677]: E1203 15:40:28.066424 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9\": container with ID starting with 48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9 not found: ID does not exist" containerID="48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9" Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.066474 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9"} err="failed to get container status \"48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9\": rpc error: code = NotFound desc = could not find container \"48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9\": container with ID starting with 48db19867172489cb602eac45178995caa51fd80ed28b179c45d1a67378218b9 not found: ID does not exist" Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.066505 4677 scope.go:117] "RemoveContainer" containerID="bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a" Dec 03 15:40:28 crc kubenswrapper[4677]: E1203 15:40:28.066927 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a\": container with ID starting with bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a not found: ID does not exist" containerID="bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a" Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.066963 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a"} err="failed to get container status \"bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a\": rpc error: code = NotFound desc = could not find container \"bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a\": container with ID starting with bd58c31a48929bbb400beb9ba0121bb84c1952e8b152ba1d9c6d59efd8b3045a not found: ID does not exist" Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.066976 4677 scope.go:117] "RemoveContainer" containerID="bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013" Dec 03 15:40:28 crc kubenswrapper[4677]: E1203 15:40:28.067330 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013\": container with ID starting with bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013 not found: ID does not exist" containerID="bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013" Dec 03 15:40:28 crc kubenswrapper[4677]: I1203 15:40:28.067370 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013"} err="failed to get container status \"bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013\": rpc error: code = NotFound desc = could not find container \"bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013\": container with ID starting with bff226e0ff64e3f2f6c08deeb600db09895e334c33f1a0bf71e6027742959013 not found: ID does not exist" Dec 03 15:40:29 crc kubenswrapper[4677]: I1203 15:40:29.990016 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:40:29 crc kubenswrapper[4677]: E1203 15:40:29.992086 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:40:29 crc kubenswrapper[4677]: I1203 15:40:29.993250 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" path="/var/lib/kubelet/pods/a16d95e4-0a07-4c8e-a023-561224f1c78c/volumes" Dec 03 15:40:42 crc kubenswrapper[4677]: I1203 15:40:42.975781 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:40:42 crc kubenswrapper[4677]: E1203 15:40:42.976731 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:40:57 crc kubenswrapper[4677]: I1203 15:40:57.976946 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:40:57 crc kubenswrapper[4677]: E1203 15:40:57.977885 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:41:10 crc kubenswrapper[4677]: I1203 15:41:10.976599 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:41:10 crc kubenswrapper[4677]: E1203 15:41:10.977449 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:41:22 crc kubenswrapper[4677]: I1203 15:41:22.976303 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:41:22 crc kubenswrapper[4677]: E1203 15:41:22.977233 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:41:36 crc kubenswrapper[4677]: I1203 15:41:36.976902 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:41:36 crc kubenswrapper[4677]: E1203 15:41:36.978695 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:41:48 crc kubenswrapper[4677]: I1203 15:41:48.976559 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:41:50 crc kubenswrapper[4677]: I1203 15:41:50.043487 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"ac8775bdd4f75d45ad0fbce0764534e98c37d9b01bef08046aa56f732a2ca8d1"} Dec 03 15:44:08 crc kubenswrapper[4677]: I1203 15:44:08.437533 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:44:08 crc kubenswrapper[4677]: I1203 15:44:08.438123 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:44:38 crc kubenswrapper[4677]: I1203 15:44:38.437331 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:44:38 crc kubenswrapper[4677]: I1203 15:44:38.437835 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.205208 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd"] Dec 03 15:45:00 crc kubenswrapper[4677]: E1203 15:45:00.206626 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="extract-utilities" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.206648 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="extract-utilities" Dec 03 15:45:00 crc kubenswrapper[4677]: E1203 15:45:00.206690 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="registry-server" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.206698 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="registry-server" Dec 03 15:45:00 crc kubenswrapper[4677]: E1203 15:45:00.206717 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="extract-content" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.206725 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="extract-content" Dec 03 15:45:00 crc kubenswrapper[4677]: E1203 15:45:00.206745 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="extract-utilities" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.206753 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="extract-utilities" Dec 03 15:45:00 crc kubenswrapper[4677]: E1203 15:45:00.206778 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="extract-content" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.206784 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="extract-content" Dec 03 15:45:00 crc kubenswrapper[4677]: E1203 15:45:00.206795 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="registry-server" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.206802 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="registry-server" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.207109 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a16d95e4-0a07-4c8e-a023-561224f1c78c" containerName="registry-server" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.207147 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="88150969-a1fa-439e-9719-ba331eb2f657" containerName="registry-server" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.207941 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.210601 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.212066 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.214389 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd"] Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.261717 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/907843b2-1251-491c-bbca-b5b355924bd1-config-volume\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.261781 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/907843b2-1251-491c-bbca-b5b355924bd1-secret-volume\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.262020 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9n4s\" (UniqueName: \"kubernetes.io/projected/907843b2-1251-491c-bbca-b5b355924bd1-kube-api-access-k9n4s\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.363654 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9n4s\" (UniqueName: \"kubernetes.io/projected/907843b2-1251-491c-bbca-b5b355924bd1-kube-api-access-k9n4s\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.363773 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/907843b2-1251-491c-bbca-b5b355924bd1-secret-volume\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.363800 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/907843b2-1251-491c-bbca-b5b355924bd1-config-volume\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.364798 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/907843b2-1251-491c-bbca-b5b355924bd1-config-volume\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.377905 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/907843b2-1251-491c-bbca-b5b355924bd1-secret-volume\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.383735 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9n4s\" (UniqueName: \"kubernetes.io/projected/907843b2-1251-491c-bbca-b5b355924bd1-kube-api-access-k9n4s\") pod \"collect-profiles-29412945-qjffd\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:00 crc kubenswrapper[4677]: I1203 15:45:00.542586 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:01 crc kubenswrapper[4677]: I1203 15:45:01.018514 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd"] Dec 03 15:45:01 crc kubenswrapper[4677]: I1203 15:45:01.129593 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" event={"ID":"907843b2-1251-491c-bbca-b5b355924bd1","Type":"ContainerStarted","Data":"dd43f6a11b3a2ee31e301718781b501f253995b771591438b4d843e57ffaf9fc"} Dec 03 15:45:02 crc kubenswrapper[4677]: I1203 15:45:02.147808 4677 generic.go:334] "Generic (PLEG): container finished" podID="907843b2-1251-491c-bbca-b5b355924bd1" containerID="3836a0e8857519156f1974b5020b42baab2eb3410d20d0999f8c8a832c593a04" exitCode=0 Dec 03 15:45:02 crc kubenswrapper[4677]: I1203 15:45:02.147870 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" event={"ID":"907843b2-1251-491c-bbca-b5b355924bd1","Type":"ContainerDied","Data":"3836a0e8857519156f1974b5020b42baab2eb3410d20d0999f8c8a832c593a04"} Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.486809 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.633143 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9n4s\" (UniqueName: \"kubernetes.io/projected/907843b2-1251-491c-bbca-b5b355924bd1-kube-api-access-k9n4s\") pod \"907843b2-1251-491c-bbca-b5b355924bd1\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.633691 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/907843b2-1251-491c-bbca-b5b355924bd1-secret-volume\") pod \"907843b2-1251-491c-bbca-b5b355924bd1\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.633871 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/907843b2-1251-491c-bbca-b5b355924bd1-config-volume\") pod \"907843b2-1251-491c-bbca-b5b355924bd1\" (UID: \"907843b2-1251-491c-bbca-b5b355924bd1\") " Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.634594 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/907843b2-1251-491c-bbca-b5b355924bd1-config-volume" (OuterVolumeSpecName: "config-volume") pod "907843b2-1251-491c-bbca-b5b355924bd1" (UID: "907843b2-1251-491c-bbca-b5b355924bd1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.634865 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/907843b2-1251-491c-bbca-b5b355924bd1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.655455 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907843b2-1251-491c-bbca-b5b355924bd1-kube-api-access-k9n4s" (OuterVolumeSpecName: "kube-api-access-k9n4s") pod "907843b2-1251-491c-bbca-b5b355924bd1" (UID: "907843b2-1251-491c-bbca-b5b355924bd1"). InnerVolumeSpecName "kube-api-access-k9n4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.657175 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907843b2-1251-491c-bbca-b5b355924bd1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "907843b2-1251-491c-bbca-b5b355924bd1" (UID: "907843b2-1251-491c-bbca-b5b355924bd1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.737281 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9n4s\" (UniqueName: \"kubernetes.io/projected/907843b2-1251-491c-bbca-b5b355924bd1-kube-api-access-k9n4s\") on node \"crc\" DevicePath \"\"" Dec 03 15:45:03 crc kubenswrapper[4677]: I1203 15:45:03.737330 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/907843b2-1251-491c-bbca-b5b355924bd1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:45:04 crc kubenswrapper[4677]: I1203 15:45:04.176690 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" event={"ID":"907843b2-1251-491c-bbca-b5b355924bd1","Type":"ContainerDied","Data":"dd43f6a11b3a2ee31e301718781b501f253995b771591438b4d843e57ffaf9fc"} Dec 03 15:45:04 crc kubenswrapper[4677]: I1203 15:45:04.176746 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd43f6a11b3a2ee31e301718781b501f253995b771591438b4d843e57ffaf9fc" Dec 03 15:45:04 crc kubenswrapper[4677]: I1203 15:45:04.177151 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412945-qjffd" Dec 03 15:45:04 crc kubenswrapper[4677]: I1203 15:45:04.573927 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x"] Dec 03 15:45:04 crc kubenswrapper[4677]: I1203 15:45:04.581675 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-sbr7x"] Dec 03 15:45:05 crc kubenswrapper[4677]: I1203 15:45:05.990221 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7b10ca4-5bf0-44c8-a83a-856deebf4170" path="/var/lib/kubelet/pods/a7b10ca4-5bf0-44c8-a83a-856deebf4170/volumes" Dec 03 15:45:08 crc kubenswrapper[4677]: I1203 15:45:08.436899 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:45:08 crc kubenswrapper[4677]: I1203 15:45:08.437771 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:45:08 crc kubenswrapper[4677]: I1203 15:45:08.437864 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:45:08 crc kubenswrapper[4677]: I1203 15:45:08.439103 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac8775bdd4f75d45ad0fbce0764534e98c37d9b01bef08046aa56f732a2ca8d1"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:45:08 crc kubenswrapper[4677]: I1203 15:45:08.439286 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://ac8775bdd4f75d45ad0fbce0764534e98c37d9b01bef08046aa56f732a2ca8d1" gracePeriod=600 Dec 03 15:45:09 crc kubenswrapper[4677]: I1203 15:45:09.231626 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="ac8775bdd4f75d45ad0fbce0764534e98c37d9b01bef08046aa56f732a2ca8d1" exitCode=0 Dec 03 15:45:09 crc kubenswrapper[4677]: I1203 15:45:09.231686 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"ac8775bdd4f75d45ad0fbce0764534e98c37d9b01bef08046aa56f732a2ca8d1"} Dec 03 15:45:09 crc kubenswrapper[4677]: I1203 15:45:09.231981 4677 scope.go:117] "RemoveContainer" containerID="b4fa4aa38129f80e8c614e0743e858f3e575901b87b1a841523b3b389b499b27" Dec 03 15:45:10 crc kubenswrapper[4677]: I1203 15:45:10.243820 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b"} Dec 03 15:45:52 crc kubenswrapper[4677]: I1203 15:45:52.715153 4677 scope.go:117] "RemoveContainer" containerID="c189a7f5fa5d0352318dee22279406cf8e8c9198cb2c292acb9ab0a30b42031d" Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.825667 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n47bq"] Dec 03 15:46:58 crc kubenswrapper[4677]: E1203 15:46:58.831368 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="907843b2-1251-491c-bbca-b5b355924bd1" containerName="collect-profiles" Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.831398 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="907843b2-1251-491c-bbca-b5b355924bd1" containerName="collect-profiles" Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.831678 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="907843b2-1251-491c-bbca-b5b355924bd1" containerName="collect-profiles" Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.833560 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.840430 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n47bq"] Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.927060 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hdn5\" (UniqueName: \"kubernetes.io/projected/ef591ae4-eba6-441f-bbf9-101b098a5a94-kube-api-access-5hdn5\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.927194 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-catalog-content\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:58 crc kubenswrapper[4677]: I1203 15:46:58.927281 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-utilities\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.028711 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-utilities\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.029490 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hdn5\" (UniqueName: \"kubernetes.io/projected/ef591ae4-eba6-441f-bbf9-101b098a5a94-kube-api-access-5hdn5\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.029652 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-catalog-content\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.030063 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-catalog-content\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.029339 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-utilities\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.052800 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hdn5\" (UniqueName: \"kubernetes.io/projected/ef591ae4-eba6-441f-bbf9-101b098a5a94-kube-api-access-5hdn5\") pod \"redhat-operators-n47bq\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.155112 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:46:59 crc kubenswrapper[4677]: I1203 15:46:59.648242 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n47bq"] Dec 03 15:47:00 crc kubenswrapper[4677]: I1203 15:47:00.418007 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerID="48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086" exitCode=0 Dec 03 15:47:00 crc kubenswrapper[4677]: I1203 15:47:00.418139 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47bq" event={"ID":"ef591ae4-eba6-441f-bbf9-101b098a5a94","Type":"ContainerDied","Data":"48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086"} Dec 03 15:47:00 crc kubenswrapper[4677]: I1203 15:47:00.418348 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47bq" event={"ID":"ef591ae4-eba6-441f-bbf9-101b098a5a94","Type":"ContainerStarted","Data":"8f928172c5eef867c05b21a4a56b9954ab06f9514d376c5d1ffe4085319b0d26"} Dec 03 15:47:00 crc kubenswrapper[4677]: I1203 15:47:00.419860 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:47:01 crc kubenswrapper[4677]: I1203 15:47:01.429762 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47bq" event={"ID":"ef591ae4-eba6-441f-bbf9-101b098a5a94","Type":"ContainerStarted","Data":"85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4"} Dec 03 15:47:03 crc kubenswrapper[4677]: I1203 15:47:03.450128 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerID="85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4" exitCode=0 Dec 03 15:47:03 crc kubenswrapper[4677]: I1203 15:47:03.450209 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47bq" event={"ID":"ef591ae4-eba6-441f-bbf9-101b098a5a94","Type":"ContainerDied","Data":"85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4"} Dec 03 15:47:05 crc kubenswrapper[4677]: I1203 15:47:05.484858 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47bq" event={"ID":"ef591ae4-eba6-441f-bbf9-101b098a5a94","Type":"ContainerStarted","Data":"b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a"} Dec 03 15:47:05 crc kubenswrapper[4677]: I1203 15:47:05.519862 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n47bq" podStartSLOduration=2.760686105 podStartE2EDuration="7.519837625s" podCreationTimestamp="2025-12-03 15:46:58 +0000 UTC" firstStartedPulling="2025-12-03 15:47:00.41956712 +0000 UTC m=+7211.165899575" lastFinishedPulling="2025-12-03 15:47:05.17871864 +0000 UTC m=+7215.925051095" observedRunningTime="2025-12-03 15:47:05.50298892 +0000 UTC m=+7216.249321395" watchObservedRunningTime="2025-12-03 15:47:05.519837625 +0000 UTC m=+7216.266170090" Dec 03 15:47:09 crc kubenswrapper[4677]: I1203 15:47:09.156105 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:47:09 crc kubenswrapper[4677]: I1203 15:47:09.156750 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:47:10 crc kubenswrapper[4677]: I1203 15:47:10.201016 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n47bq" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="registry-server" probeResult="failure" output=< Dec 03 15:47:10 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 15:47:10 crc kubenswrapper[4677]: > Dec 03 15:47:19 crc kubenswrapper[4677]: I1203 15:47:19.205683 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:47:19 crc kubenswrapper[4677]: I1203 15:47:19.258621 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:47:19 crc kubenswrapper[4677]: I1203 15:47:19.451082 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n47bq"] Dec 03 15:47:20 crc kubenswrapper[4677]: I1203 15:47:20.638742 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n47bq" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="registry-server" containerID="cri-o://b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a" gracePeriod=2 Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.612835 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.654561 4677 generic.go:334] "Generic (PLEG): container finished" podID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerID="b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a" exitCode=0 Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.654872 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47bq" event={"ID":"ef591ae4-eba6-441f-bbf9-101b098a5a94","Type":"ContainerDied","Data":"b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a"} Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.654902 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47bq" event={"ID":"ef591ae4-eba6-441f-bbf9-101b098a5a94","Type":"ContainerDied","Data":"8f928172c5eef867c05b21a4a56b9954ab06f9514d376c5d1ffe4085319b0d26"} Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.654921 4677 scope.go:117] "RemoveContainer" containerID="b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.655108 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47bq" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.678611 4677 scope.go:117] "RemoveContainer" containerID="85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.712349 4677 scope.go:117] "RemoveContainer" containerID="48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.758349 4677 scope.go:117] "RemoveContainer" containerID="b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a" Dec 03 15:47:21 crc kubenswrapper[4677]: E1203 15:47:21.759565 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a\": container with ID starting with b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a not found: ID does not exist" containerID="b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.759634 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a"} err="failed to get container status \"b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a\": rpc error: code = NotFound desc = could not find container \"b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a\": container with ID starting with b48e37e26ed6e133c7786b038480f82f7cc0b389980aab23e4c62cbd2ee2712a not found: ID does not exist" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.759659 4677 scope.go:117] "RemoveContainer" containerID="85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4" Dec 03 15:47:21 crc kubenswrapper[4677]: E1203 15:47:21.759930 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4\": container with ID starting with 85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4 not found: ID does not exist" containerID="85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.759999 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4"} err="failed to get container status \"85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4\": rpc error: code = NotFound desc = could not find container \"85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4\": container with ID starting with 85f6a507947c4c267e7fe2e2f280ceaa7c3ac1755de692014b827b1b609f43a4 not found: ID does not exist" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.760015 4677 scope.go:117] "RemoveContainer" containerID="48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086" Dec 03 15:47:21 crc kubenswrapper[4677]: E1203 15:47:21.760299 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086\": container with ID starting with 48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086 not found: ID does not exist" containerID="48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.760324 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086"} err="failed to get container status \"48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086\": rpc error: code = NotFound desc = could not find container \"48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086\": container with ID starting with 48ace3dc7f4ac7be5c6a12f1fc9f8d8a61633466423a49c745c942b231df9086 not found: ID does not exist" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.776505 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hdn5\" (UniqueName: \"kubernetes.io/projected/ef591ae4-eba6-441f-bbf9-101b098a5a94-kube-api-access-5hdn5\") pod \"ef591ae4-eba6-441f-bbf9-101b098a5a94\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.776678 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-catalog-content\") pod \"ef591ae4-eba6-441f-bbf9-101b098a5a94\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.776814 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-utilities\") pod \"ef591ae4-eba6-441f-bbf9-101b098a5a94\" (UID: \"ef591ae4-eba6-441f-bbf9-101b098a5a94\") " Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.778812 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-utilities" (OuterVolumeSpecName: "utilities") pod "ef591ae4-eba6-441f-bbf9-101b098a5a94" (UID: "ef591ae4-eba6-441f-bbf9-101b098a5a94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.783374 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef591ae4-eba6-441f-bbf9-101b098a5a94-kube-api-access-5hdn5" (OuterVolumeSpecName: "kube-api-access-5hdn5") pod "ef591ae4-eba6-441f-bbf9-101b098a5a94" (UID: "ef591ae4-eba6-441f-bbf9-101b098a5a94"). InnerVolumeSpecName "kube-api-access-5hdn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.876553 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef591ae4-eba6-441f-bbf9-101b098a5a94" (UID: "ef591ae4-eba6-441f-bbf9-101b098a5a94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.879114 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.879150 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hdn5\" (UniqueName: \"kubernetes.io/projected/ef591ae4-eba6-441f-bbf9-101b098a5a94-kube-api-access-5hdn5\") on node \"crc\" DevicePath \"\"" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.879163 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef591ae4-eba6-441f-bbf9-101b098a5a94-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:47:21 crc kubenswrapper[4677]: I1203 15:47:21.996935 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n47bq"] Dec 03 15:47:22 crc kubenswrapper[4677]: I1203 15:47:22.008994 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n47bq"] Dec 03 15:47:23 crc kubenswrapper[4677]: I1203 15:47:23.990759 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" path="/var/lib/kubelet/pods/ef591ae4-eba6-441f-bbf9-101b098a5a94/volumes" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.257642 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rfmhp/must-gather-zh684"] Dec 03 15:47:32 crc kubenswrapper[4677]: E1203 15:47:32.258552 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="extract-content" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.258566 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="extract-content" Dec 03 15:47:32 crc kubenswrapper[4677]: E1203 15:47:32.258583 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="extract-utilities" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.258589 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="extract-utilities" Dec 03 15:47:32 crc kubenswrapper[4677]: E1203 15:47:32.258625 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="registry-server" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.258630 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="registry-server" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.258832 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef591ae4-eba6-441f-bbf9-101b098a5a94" containerName="registry-server" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.259889 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.268972 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rfmhp"/"openshift-service-ca.crt" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.269310 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rfmhp"/"kube-root-ca.crt" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.287848 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rfmhp/must-gather-zh684"] Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.415983 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34e26d55-0f57-4199-8a80-b9a009649d8f-must-gather-output\") pod \"must-gather-zh684\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.416093 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcmzl\" (UniqueName: \"kubernetes.io/projected/34e26d55-0f57-4199-8a80-b9a009649d8f-kube-api-access-mcmzl\") pod \"must-gather-zh684\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.518539 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34e26d55-0f57-4199-8a80-b9a009649d8f-must-gather-output\") pod \"must-gather-zh684\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.518641 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcmzl\" (UniqueName: \"kubernetes.io/projected/34e26d55-0f57-4199-8a80-b9a009649d8f-kube-api-access-mcmzl\") pod \"must-gather-zh684\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.519012 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34e26d55-0f57-4199-8a80-b9a009649d8f-must-gather-output\") pod \"must-gather-zh684\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.544252 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcmzl\" (UniqueName: \"kubernetes.io/projected/34e26d55-0f57-4199-8a80-b9a009649d8f-kube-api-access-mcmzl\") pod \"must-gather-zh684\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:32 crc kubenswrapper[4677]: I1203 15:47:32.580524 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:47:33 crc kubenswrapper[4677]: I1203 15:47:33.075728 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rfmhp/must-gather-zh684"] Dec 03 15:47:33 crc kubenswrapper[4677]: I1203 15:47:33.797883 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/must-gather-zh684" event={"ID":"34e26d55-0f57-4199-8a80-b9a009649d8f","Type":"ContainerStarted","Data":"8d0fc5e549d21f134615677eb65223981a1573896f639e36d163a4f7a5e899fa"} Dec 03 15:47:38 crc kubenswrapper[4677]: I1203 15:47:38.437374 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:47:38 crc kubenswrapper[4677]: I1203 15:47:38.438029 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:47:42 crc kubenswrapper[4677]: I1203 15:47:42.913033 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/must-gather-zh684" event={"ID":"34e26d55-0f57-4199-8a80-b9a009649d8f","Type":"ContainerStarted","Data":"f7adac2a771fcee555002cc5232e310021957c82f9815308ed6eb3f49bf74c47"} Dec 03 15:47:48 crc kubenswrapper[4677]: I1203 15:47:48.000850 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/must-gather-zh684" event={"ID":"34e26d55-0f57-4199-8a80-b9a009649d8f","Type":"ContainerStarted","Data":"2598f87f9be583e35577869a9c7f1929190ce651ab7139ce90b6b739eb7ac519"} Dec 03 15:47:48 crc kubenswrapper[4677]: I1203 15:47:48.023753 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rfmhp/must-gather-zh684" podStartSLOduration=7.3592686050000005 podStartE2EDuration="16.023732282s" podCreationTimestamp="2025-12-03 15:47:32 +0000 UTC" firstStartedPulling="2025-12-03 15:47:33.091521587 +0000 UTC m=+7243.837854042" lastFinishedPulling="2025-12-03 15:47:41.755985264 +0000 UTC m=+7252.502317719" observedRunningTime="2025-12-03 15:47:48.016389783 +0000 UTC m=+7258.762722248" watchObservedRunningTime="2025-12-03 15:47:48.023732282 +0000 UTC m=+7258.770064737" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.832109 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-n5n5b"] Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.833921 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.836240 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rfmhp"/"default-dockercfg-nksql" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.865110 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e3d46d8d-c0f7-444d-821b-b7482ef42227-host\") pod \"crc-debug-n5n5b\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.865225 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcjmg\" (UniqueName: \"kubernetes.io/projected/e3d46d8d-c0f7-444d-821b-b7482ef42227-kube-api-access-bcjmg\") pod \"crc-debug-n5n5b\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.967436 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e3d46d8d-c0f7-444d-821b-b7482ef42227-host\") pod \"crc-debug-n5n5b\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.967796 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcjmg\" (UniqueName: \"kubernetes.io/projected/e3d46d8d-c0f7-444d-821b-b7482ef42227-kube-api-access-bcjmg\") pod \"crc-debug-n5n5b\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.967646 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e3d46d8d-c0f7-444d-821b-b7482ef42227-host\") pod \"crc-debug-n5n5b\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:51 crc kubenswrapper[4677]: I1203 15:47:51.989059 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcjmg\" (UniqueName: \"kubernetes.io/projected/e3d46d8d-c0f7-444d-821b-b7482ef42227-kube-api-access-bcjmg\") pod \"crc-debug-n5n5b\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:52 crc kubenswrapper[4677]: I1203 15:47:52.153530 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:47:52 crc kubenswrapper[4677]: W1203 15:47:52.188721 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3d46d8d_c0f7_444d_821b_b7482ef42227.slice/crio-e03b4df8b36a682c5b410d11bfe6981eef463fb646541599a1f317c156a7335d WatchSource:0}: Error finding container e03b4df8b36a682c5b410d11bfe6981eef463fb646541599a1f317c156a7335d: Status 404 returned error can't find the container with id e03b4df8b36a682c5b410d11bfe6981eef463fb646541599a1f317c156a7335d Dec 03 15:47:53 crc kubenswrapper[4677]: I1203 15:47:53.055988 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" event={"ID":"e3d46d8d-c0f7-444d-821b-b7482ef42227","Type":"ContainerStarted","Data":"e03b4df8b36a682c5b410d11bfe6981eef463fb646541599a1f317c156a7335d"} Dec 03 15:48:03 crc kubenswrapper[4677]: I1203 15:48:03.164081 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" event={"ID":"e3d46d8d-c0f7-444d-821b-b7482ef42227","Type":"ContainerStarted","Data":"49b1a78fc2af90d15918dfd3d4bfeac57f42d7262cca9b5455c8d41c7366f182"} Dec 03 15:48:03 crc kubenswrapper[4677]: I1203 15:48:03.184741 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" podStartSLOduration=1.88343284 podStartE2EDuration="12.184722424s" podCreationTimestamp="2025-12-03 15:47:51 +0000 UTC" firstStartedPulling="2025-12-03 15:47:52.190887511 +0000 UTC m=+7262.937219956" lastFinishedPulling="2025-12-03 15:48:02.492177085 +0000 UTC m=+7273.238509540" observedRunningTime="2025-12-03 15:48:03.177427386 +0000 UTC m=+7273.923759861" watchObservedRunningTime="2025-12-03 15:48:03.184722424 +0000 UTC m=+7273.931054879" Dec 03 15:48:08 crc kubenswrapper[4677]: I1203 15:48:08.437370 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:48:08 crc kubenswrapper[4677]: I1203 15:48:08.438893 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.437210 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.437768 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.437823 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.438728 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.438791 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" gracePeriod=600 Dec 03 15:48:38 crc kubenswrapper[4677]: E1203 15:48:38.575039 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.592630 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" exitCode=0 Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.592681 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b"} Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.592730 4677 scope.go:117] "RemoveContainer" containerID="ac8775bdd4f75d45ad0fbce0764534e98c37d9b01bef08046aa56f732a2ca8d1" Dec 03 15:48:38 crc kubenswrapper[4677]: I1203 15:48:38.593396 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:48:38 crc kubenswrapper[4677]: E1203 15:48:38.593681 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:48:50 crc kubenswrapper[4677]: I1203 15:48:50.755171 4677 generic.go:334] "Generic (PLEG): container finished" podID="e3d46d8d-c0f7-444d-821b-b7482ef42227" containerID="49b1a78fc2af90d15918dfd3d4bfeac57f42d7262cca9b5455c8d41c7366f182" exitCode=0 Dec 03 15:48:50 crc kubenswrapper[4677]: I1203 15:48:50.755267 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" event={"ID":"e3d46d8d-c0f7-444d-821b-b7482ef42227","Type":"ContainerDied","Data":"49b1a78fc2af90d15918dfd3d4bfeac57f42d7262cca9b5455c8d41c7366f182"} Dec 03 15:48:51 crc kubenswrapper[4677]: I1203 15:48:51.887377 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:48:51 crc kubenswrapper[4677]: I1203 15:48:51.927485 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-n5n5b"] Dec 03 15:48:51 crc kubenswrapper[4677]: I1203 15:48:51.943577 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-n5n5b"] Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.053400 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e3d46d8d-c0f7-444d-821b-b7482ef42227-host\") pod \"e3d46d8d-c0f7-444d-821b-b7482ef42227\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.053538 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3d46d8d-c0f7-444d-821b-b7482ef42227-host" (OuterVolumeSpecName: "host") pod "e3d46d8d-c0f7-444d-821b-b7482ef42227" (UID: "e3d46d8d-c0f7-444d-821b-b7482ef42227"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.054113 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcjmg\" (UniqueName: \"kubernetes.io/projected/e3d46d8d-c0f7-444d-821b-b7482ef42227-kube-api-access-bcjmg\") pod \"e3d46d8d-c0f7-444d-821b-b7482ef42227\" (UID: \"e3d46d8d-c0f7-444d-821b-b7482ef42227\") " Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.054704 4677 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e3d46d8d-c0f7-444d-821b-b7482ef42227-host\") on node \"crc\" DevicePath \"\"" Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.061116 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d46d8d-c0f7-444d-821b-b7482ef42227-kube-api-access-bcjmg" (OuterVolumeSpecName: "kube-api-access-bcjmg") pod "e3d46d8d-c0f7-444d-821b-b7482ef42227" (UID: "e3d46d8d-c0f7-444d-821b-b7482ef42227"). InnerVolumeSpecName "kube-api-access-bcjmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.156677 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcjmg\" (UniqueName: \"kubernetes.io/projected/e3d46d8d-c0f7-444d-821b-b7482ef42227-kube-api-access-bcjmg\") on node \"crc\" DevicePath \"\"" Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.786593 4677 scope.go:117] "RemoveContainer" containerID="49b1a78fc2af90d15918dfd3d4bfeac57f42d7262cca9b5455c8d41c7366f182" Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.786731 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-n5n5b" Dec 03 15:48:52 crc kubenswrapper[4677]: I1203 15:48:52.976892 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:48:52 crc kubenswrapper[4677]: E1203 15:48:52.977310 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.136159 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-rt58k"] Dec 03 15:48:53 crc kubenswrapper[4677]: E1203 15:48:53.136572 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d46d8d-c0f7-444d-821b-b7482ef42227" containerName="container-00" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.136589 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d46d8d-c0f7-444d-821b-b7482ef42227" containerName="container-00" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.136810 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d46d8d-c0f7-444d-821b-b7482ef42227" containerName="container-00" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.137488 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.143551 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rfmhp"/"default-dockercfg-nksql" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.279827 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx5hv\" (UniqueName: \"kubernetes.io/projected/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-kube-api-access-zx5hv\") pod \"crc-debug-rt58k\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.280358 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-host\") pod \"crc-debug-rt58k\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.382796 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-host\") pod \"crc-debug-rt58k\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.382991 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-host\") pod \"crc-debug-rt58k\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.383310 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx5hv\" (UniqueName: \"kubernetes.io/projected/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-kube-api-access-zx5hv\") pod \"crc-debug-rt58k\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.405401 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx5hv\" (UniqueName: \"kubernetes.io/projected/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-kube-api-access-zx5hv\") pod \"crc-debug-rt58k\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.459770 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:53 crc kubenswrapper[4677]: W1203 15:48:53.490365 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5d1bbb6_20ad_44db_9297_7e69a75a09e6.slice/crio-a58c93b90c48fd67cf7e5aa67f7b805953cc54f6347a4930f879a4da44307e28 WatchSource:0}: Error finding container a58c93b90c48fd67cf7e5aa67f7b805953cc54f6347a4930f879a4da44307e28: Status 404 returned error can't find the container with id a58c93b90c48fd67cf7e5aa67f7b805953cc54f6347a4930f879a4da44307e28 Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.795913 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" event={"ID":"f5d1bbb6-20ad-44db-9297-7e69a75a09e6","Type":"ContainerStarted","Data":"6de582ee9edefa7393a23e88815eec4c091afc4f54c5179b670507ccdf23fba7"} Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.795962 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" event={"ID":"f5d1bbb6-20ad-44db-9297-7e69a75a09e6","Type":"ContainerStarted","Data":"a58c93b90c48fd67cf7e5aa67f7b805953cc54f6347a4930f879a4da44307e28"} Dec 03 15:48:53 crc kubenswrapper[4677]: I1203 15:48:53.858834 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" podStartSLOduration=0.858814262 podStartE2EDuration="858.814262ms" podCreationTimestamp="2025-12-03 15:48:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:48:53.853850618 +0000 UTC m=+7324.600183073" watchObservedRunningTime="2025-12-03 15:48:53.858814262 +0000 UTC m=+7324.605146717" Dec 03 15:48:54 crc kubenswrapper[4677]: I1203 15:48:54.003165 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3d46d8d-c0f7-444d-821b-b7482ef42227" path="/var/lib/kubelet/pods/e3d46d8d-c0f7-444d-821b-b7482ef42227/volumes" Dec 03 15:48:54 crc kubenswrapper[4677]: I1203 15:48:54.806564 4677 generic.go:334] "Generic (PLEG): container finished" podID="f5d1bbb6-20ad-44db-9297-7e69a75a09e6" containerID="6de582ee9edefa7393a23e88815eec4c091afc4f54c5179b670507ccdf23fba7" exitCode=0 Dec 03 15:48:54 crc kubenswrapper[4677]: I1203 15:48:54.806825 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" event={"ID":"f5d1bbb6-20ad-44db-9297-7e69a75a09e6","Type":"ContainerDied","Data":"6de582ee9edefa7393a23e88815eec4c091afc4f54c5179b670507ccdf23fba7"} Dec 03 15:48:55 crc kubenswrapper[4677]: I1203 15:48:55.942951 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:55 crc kubenswrapper[4677]: I1203 15:48:55.991123 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-rt58k"] Dec 03 15:48:55 crc kubenswrapper[4677]: I1203 15:48:55.991825 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-rt58k"] Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.142529 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx5hv\" (UniqueName: \"kubernetes.io/projected/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-kube-api-access-zx5hv\") pod \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.142595 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-host\") pod \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\" (UID: \"f5d1bbb6-20ad-44db-9297-7e69a75a09e6\") " Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.144625 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-host" (OuterVolumeSpecName: "host") pod "f5d1bbb6-20ad-44db-9297-7e69a75a09e6" (UID: "f5d1bbb6-20ad-44db-9297-7e69a75a09e6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.154349 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-kube-api-access-zx5hv" (OuterVolumeSpecName: "kube-api-access-zx5hv") pod "f5d1bbb6-20ad-44db-9297-7e69a75a09e6" (UID: "f5d1bbb6-20ad-44db-9297-7e69a75a09e6"). InnerVolumeSpecName "kube-api-access-zx5hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.251960 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx5hv\" (UniqueName: \"kubernetes.io/projected/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-kube-api-access-zx5hv\") on node \"crc\" DevicePath \"\"" Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.252014 4677 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f5d1bbb6-20ad-44db-9297-7e69a75a09e6-host\") on node \"crc\" DevicePath \"\"" Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.830768 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a58c93b90c48fd67cf7e5aa67f7b805953cc54f6347a4930f879a4da44307e28" Dec 03 15:48:56 crc kubenswrapper[4677]: I1203 15:48:56.830827 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rt58k" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.181903 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-rwzc9"] Dec 03 15:48:57 crc kubenswrapper[4677]: E1203 15:48:57.182402 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d1bbb6-20ad-44db-9297-7e69a75a09e6" containerName="container-00" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.182415 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d1bbb6-20ad-44db-9297-7e69a75a09e6" containerName="container-00" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.182632 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d1bbb6-20ad-44db-9297-7e69a75a09e6" containerName="container-00" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.183370 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.191573 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rfmhp"/"default-dockercfg-nksql" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.372426 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dsnf\" (UniqueName: \"kubernetes.io/projected/db7e62f1-1349-4e35-a177-174d184db30e-kube-api-access-5dsnf\") pod \"crc-debug-rwzc9\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.372909 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/db7e62f1-1349-4e35-a177-174d184db30e-host\") pod \"crc-debug-rwzc9\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.475544 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dsnf\" (UniqueName: \"kubernetes.io/projected/db7e62f1-1349-4e35-a177-174d184db30e-kube-api-access-5dsnf\") pod \"crc-debug-rwzc9\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.475647 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/db7e62f1-1349-4e35-a177-174d184db30e-host\") pod \"crc-debug-rwzc9\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.475773 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/db7e62f1-1349-4e35-a177-174d184db30e-host\") pod \"crc-debug-rwzc9\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.495733 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dsnf\" (UniqueName: \"kubernetes.io/projected/db7e62f1-1349-4e35-a177-174d184db30e-kube-api-access-5dsnf\") pod \"crc-debug-rwzc9\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.508005 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rfmhp"/"default-dockercfg-nksql" Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.516195 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:48:57 crc kubenswrapper[4677]: W1203 15:48:57.556655 4677 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb7e62f1_1349_4e35_a177_174d184db30e.slice/crio-401718fe5f9b5f572ce5d2250b4c2f477b2ac6089e4e409a6ed5942f2b683da4 WatchSource:0}: Error finding container 401718fe5f9b5f572ce5d2250b4c2f477b2ac6089e4e409a6ed5942f2b683da4: Status 404 returned error can't find the container with id 401718fe5f9b5f572ce5d2250b4c2f477b2ac6089e4e409a6ed5942f2b683da4 Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.846239 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" event={"ID":"db7e62f1-1349-4e35-a177-174d184db30e","Type":"ContainerStarted","Data":"c2914ec2187245e60c192971cb042cbaa72ccb5e34cb4a7dfab9fb777deb22d1"} Dec 03 15:48:57 crc kubenswrapper[4677]: I1203 15:48:57.846305 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" event={"ID":"db7e62f1-1349-4e35-a177-174d184db30e","Type":"ContainerStarted","Data":"401718fe5f9b5f572ce5d2250b4c2f477b2ac6089e4e409a6ed5942f2b683da4"} Dec 03 15:48:58 crc kubenswrapper[4677]: I1203 15:48:58.000501 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d1bbb6-20ad-44db-9297-7e69a75a09e6" path="/var/lib/kubelet/pods/f5d1bbb6-20ad-44db-9297-7e69a75a09e6/volumes" Dec 03 15:48:58 crc kubenswrapper[4677]: I1203 15:48:58.857007 4677 generic.go:334] "Generic (PLEG): container finished" podID="db7e62f1-1349-4e35-a177-174d184db30e" containerID="c2914ec2187245e60c192971cb042cbaa72ccb5e34cb4a7dfab9fb777deb22d1" exitCode=0 Dec 03 15:48:58 crc kubenswrapper[4677]: I1203 15:48:58.857072 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" event={"ID":"db7e62f1-1349-4e35-a177-174d184db30e","Type":"ContainerDied","Data":"c2914ec2187245e60c192971cb042cbaa72ccb5e34cb4a7dfab9fb777deb22d1"} Dec 03 15:48:58 crc kubenswrapper[4677]: I1203 15:48:58.896162 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-rwzc9"] Dec 03 15:48:58 crc kubenswrapper[4677]: I1203 15:48:58.906274 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rfmhp/crc-debug-rwzc9"] Dec 03 15:48:59 crc kubenswrapper[4677]: I1203 15:48:59.997595 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.136812 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dsnf\" (UniqueName: \"kubernetes.io/projected/db7e62f1-1349-4e35-a177-174d184db30e-kube-api-access-5dsnf\") pod \"db7e62f1-1349-4e35-a177-174d184db30e\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.137346 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/db7e62f1-1349-4e35-a177-174d184db30e-host\") pod \"db7e62f1-1349-4e35-a177-174d184db30e\" (UID: \"db7e62f1-1349-4e35-a177-174d184db30e\") " Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.137506 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db7e62f1-1349-4e35-a177-174d184db30e-host" (OuterVolumeSpecName: "host") pod "db7e62f1-1349-4e35-a177-174d184db30e" (UID: "db7e62f1-1349-4e35-a177-174d184db30e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.138051 4677 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/db7e62f1-1349-4e35-a177-174d184db30e-host\") on node \"crc\" DevicePath \"\"" Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.141933 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7e62f1-1349-4e35-a177-174d184db30e-kube-api-access-5dsnf" (OuterVolumeSpecName: "kube-api-access-5dsnf") pod "db7e62f1-1349-4e35-a177-174d184db30e" (UID: "db7e62f1-1349-4e35-a177-174d184db30e"). InnerVolumeSpecName "kube-api-access-5dsnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.239267 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dsnf\" (UniqueName: \"kubernetes.io/projected/db7e62f1-1349-4e35-a177-174d184db30e-kube-api-access-5dsnf\") on node \"crc\" DevicePath \"\"" Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.878989 4677 scope.go:117] "RemoveContainer" containerID="c2914ec2187245e60c192971cb042cbaa72ccb5e34cb4a7dfab9fb777deb22d1" Dec 03 15:49:00 crc kubenswrapper[4677]: I1203 15:49:00.879022 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/crc-debug-rwzc9" Dec 03 15:49:01 crc kubenswrapper[4677]: I1203 15:49:01.989410 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db7e62f1-1349-4e35-a177-174d184db30e" path="/var/lib/kubelet/pods/db7e62f1-1349-4e35-a177-174d184db30e/volumes" Dec 03 15:49:07 crc kubenswrapper[4677]: I1203 15:49:07.998483 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:49:08 crc kubenswrapper[4677]: E1203 15:49:07.999299 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:49:18 crc kubenswrapper[4677]: I1203 15:49:18.978501 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:49:18 crc kubenswrapper[4677]: E1203 15:49:18.979408 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:49:26 crc kubenswrapper[4677]: I1203 15:49:26.810729 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6db8474b64-zrqnv_9c952b2e-f0e2-40a8-b5ad-16965795c194/barbican-api/0.log" Dec 03 15:49:26 crc kubenswrapper[4677]: I1203 15:49:26.912966 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6db8474b64-zrqnv_9c952b2e-f0e2-40a8-b5ad-16965795c194/barbican-api-log/0.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.067359 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67bc86b468-qwl7k_9f174682-8786-4f0e-8ce3-f72c5560bde6/barbican-keystone-listener/0.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.126731 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67bc86b468-qwl7k_9f174682-8786-4f0e-8ce3-f72c5560bde6/barbican-keystone-listener-log/0.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.283942 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64fbb7dcd5-9hc8s_e5960f5e-97d6-45c7-b597-b7ae61478585/barbican-worker/0.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.348085 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64fbb7dcd5-9hc8s_e5960f5e-97d6-45c7-b597-b7ae61478585/barbican-worker-log/0.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.382020 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rmzzb_82c6055b-a376-4795-9d0b-bd03d5b82acd/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.733840 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa84d9ae-8578-40d4-a16a-15de08868770/ceilometer-notification-agent/1.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.754620 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa84d9ae-8578-40d4-a16a-15de08868770/ceilometer-central-agent/0.log" Dec 03 15:49:27 crc kubenswrapper[4677]: I1203 15:49:27.838863 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa84d9ae-8578-40d4-a16a-15de08868770/ceilometer-central-agent/1.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.049775 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa84d9ae-8578-40d4-a16a-15de08868770/ceilometer-notification-agent/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.185460 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa84d9ae-8578-40d4-a16a-15de08868770/proxy-httpd/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.208243 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_aa84d9ae-8578-40d4-a16a-15de08868770/sg-core/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.454518 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0312747f-a587-41d5-883b-748b2b42e1ea/cinder-api-log/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.532050 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0312747f-a587-41d5-883b-748b2b42e1ea/cinder-api/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.658101 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_b663c30a-e2f1-474f-9a25-22d9ab5c571c/cinder-backup/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.769171 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_b663c30a-e2f1-474f-9a25-22d9ab5c571c/probe/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.794697 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_80f5a100-06b3-4f3c-beb1-7df2d48b6bc0/cinder-scheduler/0.log" Dec 03 15:49:28 crc kubenswrapper[4677]: I1203 15:49:28.949561 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_80f5a100-06b3-4f3c-beb1-7df2d48b6bc0/probe/0.log" Dec 03 15:49:29 crc kubenswrapper[4677]: I1203 15:49:29.139694 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_3014c1b9-a994-43e8-b490-f62b76bcf6f6/probe/0.log" Dec 03 15:49:29 crc kubenswrapper[4677]: I1203 15:49:29.163708 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_3014c1b9-a994-43e8-b490-f62b76bcf6f6/cinder-volume/0.log" Dec 03 15:49:29 crc kubenswrapper[4677]: I1203 15:49:29.392541 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_387b7e47-02b4-463b-a669-3cc5edc6fee6/cinder-volume/0.log" Dec 03 15:49:29 crc kubenswrapper[4677]: I1203 15:49:29.461082 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_387b7e47-02b4-463b-a669-3cc5edc6fee6/probe/0.log" Dec 03 15:49:29 crc kubenswrapper[4677]: I1203 15:49:29.595746 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-r4brv_85524a2c-13a3-494d-8528-9a92c581153b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:29 crc kubenswrapper[4677]: I1203 15:49:29.712575 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tcqxs_b3731b0f-16ad-4e49-8a6b-e1d20a118e86/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:29 crc kubenswrapper[4677]: I1203 15:49:29.837002 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-574fbf444c-q9lfl_60649fbd-8fcf-48d2-9327-6aef75ca6163/init/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.053868 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-574fbf444c-q9lfl_60649fbd-8fcf-48d2-9327-6aef75ca6163/init/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.219557 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-25g8q_21dd5f7c-dfc3-48bb-a771-f2dc4ce9c737/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.238223 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-574fbf444c-q9lfl_60649fbd-8fcf-48d2-9327-6aef75ca6163/dnsmasq-dns/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.411927 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c/glance-log/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.413612 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23eba9f1-c6e9-4dc3-ba5d-9cf53f3c7a9c/glance-httpd/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.586065 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8eb0a35e-349c-4fab-ab4d-c9f1d06194cc/glance-httpd/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.661280 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8eb0a35e-349c-4fab-ab4d-c9f1d06194cc/glance-log/0.log" Dec 03 15:49:30 crc kubenswrapper[4677]: I1203 15:49:30.965653 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c9d4fbd46-n2z95_0749a528-b8b2-497f-bde8-f7b7765e8068/horizon/0.log" Dec 03 15:49:31 crc kubenswrapper[4677]: I1203 15:49:31.714696 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-sxk6d_b1374bac-1ae5-47db-8fc5-736d874a8a55/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:31 crc kubenswrapper[4677]: I1203 15:49:31.726607 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-m5q8k_5c6dfb73-f46a-4233-a1bf-a22c6e9e923e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.033687 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-9ccdcc6d4-gw5r7_85993ef9-4908-479b-88bb-1a0d12b832e0/keystone-api/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.199896 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412901-55bb9_da2e0a2c-7e17-4c0c-8cee-a2e2cee92632/keystone-cron/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.316561 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-c9d4fbd46-n2z95_0749a528-b8b2-497f-bde8-f7b7765e8068/horizon-log/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.349581 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_62c96a92-e8b5-4511-8974-a40a064a8267/kube-state-metrics/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.448856 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-64ds6_891909bd-1955-436b-8664-79a41b0676e5/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.857146 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-55bcdc4f7c-89g9k_1ea332c2-0df8-470e-8985-e68ef50ed410/neutron-api/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.864269 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-55bcdc4f7c-89g9k_1ea332c2-0df8-470e-8985-e68ef50ed410/neutron-httpd/0.log" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.977355 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:49:32 crc kubenswrapper[4677]: E1203 15:49:32.977636 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:49:32 crc kubenswrapper[4677]: I1203 15:49:32.977853 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wl4px_164edf8a-8633-4592-b307-ea42bb77cc8d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:33 crc kubenswrapper[4677]: I1203 15:49:33.504829 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_fd65d357-06ac-49f7-85d0-890bba2de21d/nova-cell0-conductor-conductor/0.log" Dec 03 15:49:33 crc kubenswrapper[4677]: I1203 15:49:33.633262 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_19a36537-a4bc-4505-9025-8dfc670f6f0b/nova-api-log/0.log" Dec 03 15:49:33 crc kubenswrapper[4677]: I1203 15:49:33.919064 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1d4597ea-0a0a-4f5a-bae9-d69050c1fd56/nova-cell1-conductor-conductor/0.log" Dec 03 15:49:34 crc kubenswrapper[4677]: I1203 15:49:34.076050 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_57acac06-fa90-4fa1-bf79-f23abca7645a/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 15:49:34 crc kubenswrapper[4677]: I1203 15:49:34.212264 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hcrwl_7d24039e-bd02-41b9-86b3-1b23b50f6824/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:34 crc kubenswrapper[4677]: I1203 15:49:34.355669 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_19a36537-a4bc-4505-9025-8dfc670f6f0b/nova-api-api/0.log" Dec 03 15:49:34 crc kubenswrapper[4677]: I1203 15:49:34.368634 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9f548e7c-f339-41ce-8cb5-128b48ba2a16/nova-metadata-log/0.log" Dec 03 15:49:34 crc kubenswrapper[4677]: I1203 15:49:34.810498 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a60f05f1-4981-4841-b002-08a655ad4c66/nova-scheduler-scheduler/0.log" Dec 03 15:49:34 crc kubenswrapper[4677]: I1203 15:49:34.898416 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_072771d7-84c7-4274-8a3d-194f18a09b24/mysql-bootstrap/0.log" Dec 03 15:49:35 crc kubenswrapper[4677]: I1203 15:49:35.100350 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_072771d7-84c7-4274-8a3d-194f18a09b24/mysql-bootstrap/0.log" Dec 03 15:49:35 crc kubenswrapper[4677]: I1203 15:49:35.146795 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_072771d7-84c7-4274-8a3d-194f18a09b24/galera/0.log" Dec 03 15:49:35 crc kubenswrapper[4677]: I1203 15:49:35.363037 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a5ae3526-170f-474a-acce-41e5889470c8/mysql-bootstrap/0.log" Dec 03 15:49:35 crc kubenswrapper[4677]: I1203 15:49:35.579245 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a5ae3526-170f-474a-acce-41e5889470c8/galera/0.log" Dec 03 15:49:35 crc kubenswrapper[4677]: I1203 15:49:35.587708 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a5ae3526-170f-474a-acce-41e5889470c8/mysql-bootstrap/0.log" Dec 03 15:49:35 crc kubenswrapper[4677]: I1203 15:49:35.841432 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_63acf6b9-c8d3-4e57-a08c-63dff049e551/openstackclient/0.log" Dec 03 15:49:35 crc kubenswrapper[4677]: I1203 15:49:35.983515 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jpkmm_f14465fa-42ae-4e85-bb6f-f2ef1bc5b681/ovn-controller/0.log" Dec 03 15:49:36 crc kubenswrapper[4677]: I1203 15:49:36.099772 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-bfpgf_08873f8c-1b49-42cf-914c-150fd7ec7b09/openstack-network-exporter/0.log" Dec 03 15:49:36 crc kubenswrapper[4677]: I1203 15:49:36.567681 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-msfg2_d11b3a89-22f0-4768-bc37-701a49737694/ovsdb-server-init/0.log" Dec 03 15:49:36 crc kubenswrapper[4677]: I1203 15:49:36.864586 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-msfg2_d11b3a89-22f0-4768-bc37-701a49737694/ovsdb-server/0.log" Dec 03 15:49:36 crc kubenswrapper[4677]: I1203 15:49:36.865625 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-msfg2_d11b3a89-22f0-4768-bc37-701a49737694/ovsdb-server-init/0.log" Dec 03 15:49:36 crc kubenswrapper[4677]: I1203 15:49:36.888588 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-msfg2_d11b3a89-22f0-4768-bc37-701a49737694/ovs-vswitchd/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.192835 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nr8kf_aae40ead-5b0e-4308-9b3c-fe2ef3cc8eb6/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.387592 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a232e198-17b4-4617-a664-b8db60b0d3fe/ovn-northd/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.403530 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a232e198-17b4-4617-a664-b8db60b0d3fe/openstack-network-exporter/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.515109 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9f548e7c-f339-41ce-8cb5-128b48ba2a16/nova-metadata-metadata/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.655303 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_62024578-ee41-4839-9846-9a76d8785d6e/ovsdbserver-nb/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.688204 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_62024578-ee41-4839-9846-9a76d8785d6e/openstack-network-exporter/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.897078 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9b7e7e94-5f47-4963-acb0-f40d4ae3ca16/ovsdbserver-sb/0.log" Dec 03 15:49:37 crc kubenswrapper[4677]: I1203 15:49:37.920539 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9b7e7e94-5f47-4963-acb0-f40d4ae3ca16/openstack-network-exporter/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.370146 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-67bf5cb8b6-4lwks_372640de-08f9-4020-9ccf-294678119877/placement-api/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.390747 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_429e5aed-7233-4056-a808-9afc77fa5011/init-config-reloader/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.412082 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-67bf5cb8b6-4lwks_372640de-08f9-4020-9ccf-294678119877/placement-log/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.660145 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_429e5aed-7233-4056-a808-9afc77fa5011/init-config-reloader/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.679247 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_429e5aed-7233-4056-a808-9afc77fa5011/thanos-sidecar/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.686632 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_429e5aed-7233-4056-a808-9afc77fa5011/config-reloader/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.748766 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_429e5aed-7233-4056-a808-9afc77fa5011/prometheus/0.log" Dec 03 15:49:38 crc kubenswrapper[4677]: I1203 15:49:38.970739 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_29df9745-7bd4-428e-9714-624cf6f66d7e/setup-container/0.log" Dec 03 15:49:39 crc kubenswrapper[4677]: I1203 15:49:39.188734 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_29df9745-7bd4-428e-9714-624cf6f66d7e/setup-container/0.log" Dec 03 15:49:39 crc kubenswrapper[4677]: I1203 15:49:39.276526 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_315da922-99be-44c7-81d8-24f0cf55490a/setup-container/0.log" Dec 03 15:49:39 crc kubenswrapper[4677]: I1203 15:49:39.280189 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_29df9745-7bd4-428e-9714-624cf6f66d7e/rabbitmq/0.log" Dec 03 15:49:39 crc kubenswrapper[4677]: I1203 15:49:39.573899 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_315da922-99be-44c7-81d8-24f0cf55490a/setup-container/0.log" Dec 03 15:49:39 crc kubenswrapper[4677]: I1203 15:49:39.703063 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_315da922-99be-44c7-81d8-24f0cf55490a/rabbitmq/0.log" Dec 03 15:49:39 crc kubenswrapper[4677]: I1203 15:49:39.727885 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52fb034c-06d4-4cb2-be6a-0c49a521b0aa/setup-container/0.log" Dec 03 15:49:39 crc kubenswrapper[4677]: I1203 15:49:39.921094 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52fb034c-06d4-4cb2-be6a-0c49a521b0aa/setup-container/0.log" Dec 03 15:49:40 crc kubenswrapper[4677]: I1203 15:49:40.072701 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52fb034c-06d4-4cb2-be6a-0c49a521b0aa/rabbitmq/0.log" Dec 03 15:49:40 crc kubenswrapper[4677]: I1203 15:49:40.095976 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zmzgg_054f88ba-5afa-48e7-89c5-0573ae4dc14b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:40 crc kubenswrapper[4677]: I1203 15:49:40.363985 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-8dplp_c24c4c0c-e7bc-44d1-8f09-99422afc8927/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:40 crc kubenswrapper[4677]: I1203 15:49:40.445747 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-d55sv_c2dc4bf3-fecf-4b2a-a664-d5c543196286/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:40 crc kubenswrapper[4677]: I1203 15:49:40.636837 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kx6wh_5f602efb-8c0c-4ba7-a297-d24706833f00/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:40 crc kubenswrapper[4677]: I1203 15:49:40.720812 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-pwcrc_da805608-f665-4c50-baf6-b90b94c6fd2c/ssh-known-hosts-edpm-deployment/0.log" Dec 03 15:49:40 crc kubenswrapper[4677]: I1203 15:49:40.952590 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-9cc645f55-v5xrn_5da71451-5801-4ce9-b177-cd847b635e24/proxy-server/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.081828 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-qqscx_d39bf749-8b60-4833-97e0-b00791e11467/swift-ring-rebalance/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.176380 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-9cc645f55-v5xrn_5da71451-5801-4ce9-b177-cd847b635e24/proxy-httpd/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.583677 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/account-auditor/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.596524 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/account-reaper/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.727278 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/account-replicator/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.738719 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/account-server/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.763922 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_79a2ad48-62e8-4bc2-a569-b110da1c20ea/memcached/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.783397 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/container-auditor/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.904180 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/container-replicator/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.935641 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/container-server/0.log" Dec 03 15:49:41 crc kubenswrapper[4677]: I1203 15:49:41.968716 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/container-updater/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.024886 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/object-auditor/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.050680 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/object-expirer/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.146239 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/object-server/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.200586 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/object-replicator/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.231108 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/object-updater/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.277260 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/rsync/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.295307 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b25cf219-67f3-4764-a5da-3dbe89f71641/swift-recon-cron/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.442207 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s6c97_34703414-a555-4edf-addc-8c36c829718b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:42 crc kubenswrapper[4677]: I1203 15:49:42.617839 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-cj4kr_aa53aac3-038e-4728-bbec-afce4e5c7509/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:49:43 crc kubenswrapper[4677]: I1203 15:49:43.310650 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_e4245f29-1973-483f-ad9c-ab450ae9f6d6/tempest-tests-tempest-tests-runner/0.log" Dec 03 15:49:43 crc kubenswrapper[4677]: I1203 15:49:43.368152 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_2c0001c6-665a-4888-a2ff-02819272545c/watcher-applier/0.log" Dec 03 15:49:45 crc kubenswrapper[4677]: I1203 15:49:45.492316 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_12d8e83b-2cf0-4731-9c86-6a75651386b8/watcher-api-log/0.log" Dec 03 15:49:45 crc kubenswrapper[4677]: I1203 15:49:45.650838 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_8ed2ece8-e640-495d-b5fb-b0cd73a190e2/watcher-decision-engine/0.log" Dec 03 15:49:46 crc kubenswrapper[4677]: I1203 15:49:46.975768 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:49:46 crc kubenswrapper[4677]: E1203 15:49:46.977296 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:49:48 crc kubenswrapper[4677]: I1203 15:49:48.409889 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_12d8e83b-2cf0-4731-9c86-6a75651386b8/watcher-api/0.log" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.321786 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6c9qs"] Dec 03 15:49:59 crc kubenswrapper[4677]: E1203 15:49:59.323482 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7e62f1-1349-4e35-a177-174d184db30e" containerName="container-00" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.323563 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7e62f1-1349-4e35-a177-174d184db30e" containerName="container-00" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.323903 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7e62f1-1349-4e35-a177-174d184db30e" containerName="container-00" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.331047 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.342448 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6c9qs"] Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.444161 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-utilities\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.444356 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hxhk\" (UniqueName: \"kubernetes.io/projected/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-kube-api-access-5hxhk\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.444447 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-catalog-content\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.546457 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-catalog-content\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.546565 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-utilities\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.546722 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hxhk\" (UniqueName: \"kubernetes.io/projected/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-kube-api-access-5hxhk\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.547527 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-catalog-content\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.547745 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-utilities\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.576092 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hxhk\" (UniqueName: \"kubernetes.io/projected/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-kube-api-access-5hxhk\") pod \"redhat-marketplace-6c9qs\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:49:59 crc kubenswrapper[4677]: I1203 15:49:59.654142 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:50:00 crc kubenswrapper[4677]: I1203 15:50:00.356243 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6c9qs"] Dec 03 15:50:00 crc kubenswrapper[4677]: I1203 15:50:00.518074 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6c9qs" event={"ID":"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022","Type":"ContainerStarted","Data":"6e021658e07d4fe321e189bf89948d692eca23d7d05173c98cbce76befa8eec2"} Dec 03 15:50:00 crc kubenswrapper[4677]: I1203 15:50:00.977512 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:50:00 crc kubenswrapper[4677]: E1203 15:50:00.978105 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:50:01 crc kubenswrapper[4677]: I1203 15:50:01.533612 4677 generic.go:334] "Generic (PLEG): container finished" podID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerID="07cbf497ab05a429942baa0e33d8b25df96784e35a48d0ca7464b5adb1297f7c" exitCode=0 Dec 03 15:50:01 crc kubenswrapper[4677]: I1203 15:50:01.533657 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6c9qs" event={"ID":"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022","Type":"ContainerDied","Data":"07cbf497ab05a429942baa0e33d8b25df96784e35a48d0ca7464b5adb1297f7c"} Dec 03 15:50:04 crc kubenswrapper[4677]: I1203 15:50:04.562049 4677 generic.go:334] "Generic (PLEG): container finished" podID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerID="5cbda238dec9e25996810eee6d90c0df054f6cb73e947bcb48bfcd9aff09e71d" exitCode=0 Dec 03 15:50:04 crc kubenswrapper[4677]: I1203 15:50:04.562149 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6c9qs" event={"ID":"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022","Type":"ContainerDied","Data":"5cbda238dec9e25996810eee6d90c0df054f6cb73e947bcb48bfcd9aff09e71d"} Dec 03 15:50:05 crc kubenswrapper[4677]: I1203 15:50:05.576534 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6c9qs" event={"ID":"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022","Type":"ContainerStarted","Data":"42e1c1836329967c90cb6a891d43cee2479b3740093563038c26e0520e6cd9ab"} Dec 03 15:50:05 crc kubenswrapper[4677]: I1203 15:50:05.601287 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6c9qs" podStartSLOduration=3.181205056 podStartE2EDuration="6.601265017s" podCreationTimestamp="2025-12-03 15:49:59 +0000 UTC" firstStartedPulling="2025-12-03 15:50:01.536511194 +0000 UTC m=+7392.282843659" lastFinishedPulling="2025-12-03 15:50:04.956571165 +0000 UTC m=+7395.702903620" observedRunningTime="2025-12-03 15:50:05.594065604 +0000 UTC m=+7396.340398089" watchObservedRunningTime="2025-12-03 15:50:05.601265017 +0000 UTC m=+7396.347597482" Dec 03 15:50:09 crc kubenswrapper[4677]: I1203 15:50:09.654891 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:50:09 crc kubenswrapper[4677]: I1203 15:50:09.655293 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:50:09 crc kubenswrapper[4677]: I1203 15:50:09.712099 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:50:10 crc kubenswrapper[4677]: I1203 15:50:10.692478 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:50:10 crc kubenswrapper[4677]: I1203 15:50:10.744859 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6c9qs"] Dec 03 15:50:10 crc kubenswrapper[4677]: I1203 15:50:10.867206 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj_f962a042-cf75-46fe-8fb1-3c04aaa8043b/util/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.079662 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj_f962a042-cf75-46fe-8fb1-3c04aaa8043b/util/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.158111 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj_f962a042-cf75-46fe-8fb1-3c04aaa8043b/pull/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.191211 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj_f962a042-cf75-46fe-8fb1-3c04aaa8043b/pull/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.389728 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj_f962a042-cf75-46fe-8fb1-3c04aaa8043b/util/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.415982 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj_f962a042-cf75-46fe-8fb1-3c04aaa8043b/pull/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.435316 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6c95319116f562471a7215f64a3853fcd870f65555714d955da0eb8f936hgqj_f962a042-cf75-46fe-8fb1-3c04aaa8043b/extract/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.577855 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jxph7_df1820a3-7030-449d-ad5d-c8bbdb75ad28/kube-rbac-proxy/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.586469 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jxph7_df1820a3-7030-449d-ad5d-c8bbdb75ad28/manager/1.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.676692 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jxph7_df1820a3-7030-449d-ad5d-c8bbdb75ad28/manager/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.796754 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-x2q6z_460e4110-77d5-476e-88d3-d9ccec539f98/kube-rbac-proxy/0.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.820702 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-x2q6z_460e4110-77d5-476e-88d3-d9ccec539f98/manager/1.log" Dec 03 15:50:11 crc kubenswrapper[4677]: I1203 15:50:11.940178 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-x2q6z_460e4110-77d5-476e-88d3-d9ccec539f98/manager/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.032058 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-45txl_5236b4d8-b675-4e53-9c7c-c33606436dff/kube-rbac-proxy/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.103262 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-45txl_5236b4d8-b675-4e53-9c7c-c33606436dff/manager/1.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.149304 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-45txl_5236b4d8-b675-4e53-9c7c-c33606436dff/manager/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.260195 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-8rll8_226bed84-3bd3-4f3b-ad1a-ab8030a71b34/kube-rbac-proxy/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.329652 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-8rll8_226bed84-3bd3-4f3b-ad1a-ab8030a71b34/manager/1.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.423597 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-8rll8_226bed84-3bd3-4f3b-ad1a-ab8030a71b34/manager/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.487141 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-vr9nf_8d9537f6-e3da-4acb-af76-4ad3bbc403ab/kube-rbac-proxy/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.596031 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-vr9nf_8d9537f6-e3da-4acb-af76-4ad3bbc403ab/manager/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.635409 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-vr9nf_8d9537f6-e3da-4acb-af76-4ad3bbc403ab/manager/1.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.648089 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6c9qs" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="registry-server" containerID="cri-o://42e1c1836329967c90cb6a891d43cee2479b3740093563038c26e0520e6cd9ab" gracePeriod=2 Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.709562 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kqdd2_01e63cef-045a-42e0-8776-2f07b1187a40/kube-rbac-proxy/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.886336 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kqdd2_01e63cef-045a-42e0-8776-2f07b1187a40/manager/0.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.896012 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kqdd2_01e63cef-045a-42e0-8776-2f07b1187a40/manager/1.log" Dec 03 15:50:12 crc kubenswrapper[4677]: I1203 15:50:12.919672 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-qhdx7_5b6f5612-8a33-46f2-8f4e-415d5bc2e807/kube-rbac-proxy/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.107748 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-qhdx7_5b6f5612-8a33-46f2-8f4e-415d5bc2e807/manager/1.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.163711 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-rcgbn_cca481a1-d874-446a-9f63-926247653192/kube-rbac-proxy/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.269749 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-qhdx7_5b6f5612-8a33-46f2-8f4e-415d5bc2e807/manager/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.296721 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-rcgbn_cca481a1-d874-446a-9f63-926247653192/manager/1.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.408574 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-rcgbn_cca481a1-d874-446a-9f63-926247653192/manager/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.491625 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-5f5nf_af346d56-cdf6-408d-87fc-6431b12a9cd8/kube-rbac-proxy/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.502255 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-5f5nf_af346d56-cdf6-408d-87fc-6431b12a9cd8/manager/1.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.673690 4677 generic.go:334] "Generic (PLEG): container finished" podID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerID="42e1c1836329967c90cb6a891d43cee2479b3740093563038c26e0520e6cd9ab" exitCode=0 Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.673783 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6c9qs" event={"ID":"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022","Type":"ContainerDied","Data":"42e1c1836329967c90cb6a891d43cee2479b3740093563038c26e0520e6cd9ab"} Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.740991 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-l2wcn_829c01d5-d4d0-40a4-9513-0e60b322d4d0/manager/1.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.805525 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-5f5nf_af346d56-cdf6-408d-87fc-6431b12a9cd8/manager/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.816379 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-l2wcn_829c01d5-d4d0-40a4-9513-0e60b322d4d0/kube-rbac-proxy/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.856711 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.941422 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-l2wcn_829c01d5-d4d0-40a4-9513-0e60b322d4d0/manager/0.log" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.960807 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-utilities\") pod \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.961032 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-catalog-content\") pod \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.961159 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hxhk\" (UniqueName: \"kubernetes.io/projected/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-kube-api-access-5hxhk\") pod \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\" (UID: \"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022\") " Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.962924 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-utilities" (OuterVolumeSpecName: "utilities") pod "6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" (UID: "6ac9177d-9aff-4f77-91f8-d7cc9ab7a022"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.967791 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-kube-api-access-5hxhk" (OuterVolumeSpecName: "kube-api-access-5hxhk") pod "6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" (UID: "6ac9177d-9aff-4f77-91f8-d7cc9ab7a022"). InnerVolumeSpecName "kube-api-access-5hxhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:50:13 crc kubenswrapper[4677]: I1203 15:50:13.986244 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" (UID: "6ac9177d-9aff-4f77-91f8-d7cc9ab7a022"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.063725 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.063761 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hxhk\" (UniqueName: \"kubernetes.io/projected/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-kube-api-access-5hxhk\") on node \"crc\" DevicePath \"\"" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.063771 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.090364 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-4tw2x_1e0b7fb1-cdb4-44a9-9508-4939f1038023/kube-rbac-proxy/0.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.125567 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-4tw2x_1e0b7fb1-cdb4-44a9-9508-4939f1038023/manager/1.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.303100 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-4tw2x_1e0b7fb1-cdb4-44a9-9508-4939f1038023/manager/0.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.350532 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-n6xjd_0ffc0c32-9591-4a0d-b7fb-10b8faa85c18/kube-rbac-proxy/0.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.409787 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-n6xjd_0ffc0c32-9591-4a0d-b7fb-10b8faa85c18/manager/1.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.543582 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vk8s9_7270260d-7534-4b53-be80-69fbee93fdb7/kube-rbac-proxy/0.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.581096 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-n6xjd_0ffc0c32-9591-4a0d-b7fb-10b8faa85c18/manager/0.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.614896 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vk8s9_7270260d-7534-4b53-be80-69fbee93fdb7/manager/1.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.691472 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6c9qs" event={"ID":"6ac9177d-9aff-4f77-91f8-d7cc9ab7a022","Type":"ContainerDied","Data":"6e021658e07d4fe321e189bf89948d692eca23d7d05173c98cbce76befa8eec2"} Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.691529 4677 scope.go:117] "RemoveContainer" containerID="42e1c1836329967c90cb6a891d43cee2479b3740093563038c26e0520e6cd9ab" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.691561 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6c9qs" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.718664 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6c9qs"] Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.719898 4677 scope.go:117] "RemoveContainer" containerID="5cbda238dec9e25996810eee6d90c0df054f6cb73e947bcb48bfcd9aff09e71d" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.752208 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6c9qs"] Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.762388 4677 scope.go:117] "RemoveContainer" containerID="07cbf497ab05a429942baa0e33d8b25df96784e35a48d0ca7464b5adb1297f7c" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.850631 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-vv4wp_59757e80-419c-40f9-9ab2-f6dbeff58f7c/kube-rbac-proxy/0.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.860313 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-vk8s9_7270260d-7534-4b53-be80-69fbee93fdb7/manager/0.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.878074 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-vv4wp_59757e80-419c-40f9-9ab2-f6dbeff58f7c/manager/1.log" Dec 03 15:50:14 crc kubenswrapper[4677]: I1203 15:50:14.976745 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:50:14 crc kubenswrapper[4677]: E1203 15:50:14.977199 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.034130 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-vv4wp_59757e80-419c-40f9-9ab2-f6dbeff58f7c/manager/0.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.143133 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk_de88ce12-43f1-4e18-ad73-f7d2c222c4ca/manager/1.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.426532 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk_de88ce12-43f1-4e18-ad73-f7d2c222c4ca/kube-rbac-proxy/0.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.426681 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4dshfk_de88ce12-43f1-4e18-ad73-f7d2c222c4ca/manager/0.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.454226 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-775cffb9d6-lk7pd_f5d376fc-835d-44d7-954a-f838a22f7aa8/manager/1.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.646981 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b44f76b7c-m9c5x_6c72cd7c-f759-4346-a975-4f1cbec9cbf7/operator/1.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.695107 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-t5d47_d21b8d21-356b-4cb9-ab6e-1a08320686fc/registry-server/0.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.871826 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-hfwgd_da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6/kube-rbac-proxy/0.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.910192 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b44f76b7c-m9c5x_6c72cd7c-f759-4346-a975-4f1cbec9cbf7/operator/0.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.942421 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-hfwgd_da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6/manager/1.log" Dec 03 15:50:15 crc kubenswrapper[4677]: I1203 15:50:15.989578 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" path="/var/lib/kubelet/pods/6ac9177d-9aff-4f77-91f8-d7cc9ab7a022/volumes" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.122076 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-qxf8w_47d9316a-afea-441d-8c42-45e90efe4d5a/kube-rbac-proxy/0.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.234431 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-qxf8w_47d9316a-afea-441d-8c42-45e90efe4d5a/manager/1.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.283267 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-hfwgd_da22d3ae-8680-4ca0-a3ec-be3ee2c41bc6/manager/0.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.376123 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-qxf8w_47d9316a-afea-441d-8c42-45e90efe4d5a/manager/0.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.537417 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qnqg7_27ee4b91-853f-444b-8663-d913a1d9d1e5/operator/1.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.559566 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qnqg7_27ee4b91-853f-444b-8663-d913a1d9d1e5/operator/0.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.746679 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-8jjpl_bbb9ba89-c82a-4cb6-8346-a76d2f24fd38/kube-rbac-proxy/0.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.755357 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-8jjpl_bbb9ba89-c82a-4cb6-8346-a76d2f24fd38/manager/1.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.858050 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-8jjpl_bbb9ba89-c82a-4cb6-8346-a76d2f24fd38/manager/0.log" Dec 03 15:50:16 crc kubenswrapper[4677]: I1203 15:50:16.919206 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-9jc84_882195ac-f224-4563-b00d-0291e40f3204/kube-rbac-proxy/0.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.042405 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-9jc84_882195ac-f224-4563-b00d-0291e40f3204/manager/1.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.193358 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-jmxdr_85cba130-9ba2-48b7-917f-14946bc72419/kube-rbac-proxy/0.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.195165 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-775cffb9d6-lk7pd_f5d376fc-835d-44d7-954a-f838a22f7aa8/manager/0.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.317002 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-jmxdr_85cba130-9ba2-48b7-917f-14946bc72419/manager/1.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.373154 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-9jc84_882195ac-f224-4563-b00d-0291e40f3204/manager/0.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.474483 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6757bc7c76-rqljj_40fcffeb-3d82-431f-a208-c06bc3f30557/kube-rbac-proxy/0.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.498944 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-jmxdr_85cba130-9ba2-48b7-917f-14946bc72419/manager/0.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.703347 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6757bc7c76-rqljj_40fcffeb-3d82-431f-a208-c06bc3f30557/manager/1.log" Dec 03 15:50:17 crc kubenswrapper[4677]: I1203 15:50:17.860635 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6757bc7c76-rqljj_40fcffeb-3d82-431f-a208-c06bc3f30557/manager/0.log" Dec 03 15:50:27 crc kubenswrapper[4677]: I1203 15:50:27.976046 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:50:27 crc kubenswrapper[4677]: E1203 15:50:27.976984 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:50:37 crc kubenswrapper[4677]: I1203 15:50:37.274583 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2c9gq_da040c14-6dff-4426-8258-6bd0e3e14cb6/control-plane-machine-set-operator/0.log" Dec 03 15:50:37 crc kubenswrapper[4677]: I1203 15:50:37.450289 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d4p7j_93a0fb86-4f19-4023-b777-c58381af5047/kube-rbac-proxy/0.log" Dec 03 15:50:37 crc kubenswrapper[4677]: I1203 15:50:37.521558 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d4p7j_93a0fb86-4f19-4023-b777-c58381af5047/machine-api-operator/0.log" Dec 03 15:50:42 crc kubenswrapper[4677]: I1203 15:50:42.976041 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:50:42 crc kubenswrapper[4677]: E1203 15:50:42.976977 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.036064 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-46whr"] Dec 03 15:50:45 crc kubenswrapper[4677]: E1203 15:50:45.037130 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="extract-content" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.037154 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="extract-content" Dec 03 15:50:45 crc kubenswrapper[4677]: E1203 15:50:45.037194 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="extract-utilities" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.037203 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="extract-utilities" Dec 03 15:50:45 crc kubenswrapper[4677]: E1203 15:50:45.037228 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="registry-server" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.037236 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="registry-server" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.037501 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac9177d-9aff-4f77-91f8-d7cc9ab7a022" containerName="registry-server" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.039387 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.052634 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-46whr"] Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.218812 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-catalog-content\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.218894 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs74p\" (UniqueName: \"kubernetes.io/projected/ff5aa690-9273-42de-a62a-028235952ad2-kube-api-access-fs74p\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.219023 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-utilities\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.320879 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-catalog-content\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.321531 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-catalog-content\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.321715 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs74p\" (UniqueName: \"kubernetes.io/projected/ff5aa690-9273-42de-a62a-028235952ad2-kube-api-access-fs74p\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.321833 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-utilities\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.322181 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-utilities\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.345869 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs74p\" (UniqueName: \"kubernetes.io/projected/ff5aa690-9273-42de-a62a-028235952ad2-kube-api-access-fs74p\") pod \"certified-operators-46whr\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:45 crc kubenswrapper[4677]: I1203 15:50:45.366687 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:46 crc kubenswrapper[4677]: I1203 15:50:46.027098 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-46whr"] Dec 03 15:50:47 crc kubenswrapper[4677]: I1203 15:50:47.019086 4677 generic.go:334] "Generic (PLEG): container finished" podID="ff5aa690-9273-42de-a62a-028235952ad2" containerID="72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4" exitCode=0 Dec 03 15:50:47 crc kubenswrapper[4677]: I1203 15:50:47.019196 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46whr" event={"ID":"ff5aa690-9273-42de-a62a-028235952ad2","Type":"ContainerDied","Data":"72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4"} Dec 03 15:50:47 crc kubenswrapper[4677]: I1203 15:50:47.019714 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46whr" event={"ID":"ff5aa690-9273-42de-a62a-028235952ad2","Type":"ContainerStarted","Data":"cbe812794ae30d9edc792348f35332fdfe34fc35ae605fbef4ca2783e62690fc"} Dec 03 15:50:48 crc kubenswrapper[4677]: I1203 15:50:48.033647 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46whr" event={"ID":"ff5aa690-9273-42de-a62a-028235952ad2","Type":"ContainerStarted","Data":"54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13"} Dec 03 15:50:49 crc kubenswrapper[4677]: I1203 15:50:49.043665 4677 generic.go:334] "Generic (PLEG): container finished" podID="ff5aa690-9273-42de-a62a-028235952ad2" containerID="54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13" exitCode=0 Dec 03 15:50:49 crc kubenswrapper[4677]: I1203 15:50:49.043757 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46whr" event={"ID":"ff5aa690-9273-42de-a62a-028235952ad2","Type":"ContainerDied","Data":"54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13"} Dec 03 15:50:50 crc kubenswrapper[4677]: I1203 15:50:50.056826 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46whr" event={"ID":"ff5aa690-9273-42de-a62a-028235952ad2","Type":"ContainerStarted","Data":"d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a"} Dec 03 15:50:50 crc kubenswrapper[4677]: I1203 15:50:50.077395 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-46whr" podStartSLOduration=2.6953374549999998 podStartE2EDuration="5.077363495s" podCreationTimestamp="2025-12-03 15:50:45 +0000 UTC" firstStartedPulling="2025-12-03 15:50:47.021261213 +0000 UTC m=+7437.767593668" lastFinishedPulling="2025-12-03 15:50:49.403287253 +0000 UTC m=+7440.149619708" observedRunningTime="2025-12-03 15:50:50.074410916 +0000 UTC m=+7440.820743391" watchObservedRunningTime="2025-12-03 15:50:50.077363495 +0000 UTC m=+7440.823695950" Dec 03 15:50:51 crc kubenswrapper[4677]: I1203 15:50:51.015281 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-b5kfm_ccc13e6c-e114-482c-8882-fed3fd6324a0/cert-manager-controller/1.log" Dec 03 15:50:51 crc kubenswrapper[4677]: I1203 15:50:51.107718 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-b5kfm_ccc13e6c-e114-482c-8882-fed3fd6324a0/cert-manager-controller/0.log" Dec 03 15:50:51 crc kubenswrapper[4677]: I1203 15:50:51.316544 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pwkvm_84174520-f102-4fcb-ae53-bb5dd2218549/cert-manager-cainjector/0.log" Dec 03 15:50:51 crc kubenswrapper[4677]: I1203 15:50:51.366598 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-zbfln_ecdcc3d2-b371-4439-93b8-fe9e6a945ed9/cert-manager-webhook/0.log" Dec 03 15:50:55 crc kubenswrapper[4677]: I1203 15:50:55.367346 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:55 crc kubenswrapper[4677]: I1203 15:50:55.369076 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:55 crc kubenswrapper[4677]: I1203 15:50:55.416909 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:56 crc kubenswrapper[4677]: I1203 15:50:56.159242 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:56 crc kubenswrapper[4677]: I1203 15:50:56.210259 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-46whr"] Dec 03 15:50:56 crc kubenswrapper[4677]: I1203 15:50:56.976844 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:50:56 crc kubenswrapper[4677]: E1203 15:50:56.977336 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.135524 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-46whr" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="registry-server" containerID="cri-o://d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a" gracePeriod=2 Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.646724 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.718603 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-utilities\") pod \"ff5aa690-9273-42de-a62a-028235952ad2\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.718913 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-catalog-content\") pod \"ff5aa690-9273-42de-a62a-028235952ad2\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.719050 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs74p\" (UniqueName: \"kubernetes.io/projected/ff5aa690-9273-42de-a62a-028235952ad2-kube-api-access-fs74p\") pod \"ff5aa690-9273-42de-a62a-028235952ad2\" (UID: \"ff5aa690-9273-42de-a62a-028235952ad2\") " Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.719488 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-utilities" (OuterVolumeSpecName: "utilities") pod "ff5aa690-9273-42de-a62a-028235952ad2" (UID: "ff5aa690-9273-42de-a62a-028235952ad2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.719693 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.726575 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5aa690-9273-42de-a62a-028235952ad2-kube-api-access-fs74p" (OuterVolumeSpecName: "kube-api-access-fs74p") pod "ff5aa690-9273-42de-a62a-028235952ad2" (UID: "ff5aa690-9273-42de-a62a-028235952ad2"). InnerVolumeSpecName "kube-api-access-fs74p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.807797 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff5aa690-9273-42de-a62a-028235952ad2" (UID: "ff5aa690-9273-42de-a62a-028235952ad2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.820704 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5aa690-9273-42de-a62a-028235952ad2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:50:58 crc kubenswrapper[4677]: I1203 15:50:58.820747 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs74p\" (UniqueName: \"kubernetes.io/projected/ff5aa690-9273-42de-a62a-028235952ad2-kube-api-access-fs74p\") on node \"crc\" DevicePath \"\"" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.153764 4677 generic.go:334] "Generic (PLEG): container finished" podID="ff5aa690-9273-42de-a62a-028235952ad2" containerID="d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a" exitCode=0 Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.153851 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46whr" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.153843 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46whr" event={"ID":"ff5aa690-9273-42de-a62a-028235952ad2","Type":"ContainerDied","Data":"d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a"} Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.155330 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46whr" event={"ID":"ff5aa690-9273-42de-a62a-028235952ad2","Type":"ContainerDied","Data":"cbe812794ae30d9edc792348f35332fdfe34fc35ae605fbef4ca2783e62690fc"} Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.155365 4677 scope.go:117] "RemoveContainer" containerID="d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.190402 4677 scope.go:117] "RemoveContainer" containerID="54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.218258 4677 scope.go:117] "RemoveContainer" containerID="72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.221056 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-46whr"] Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.230015 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-46whr"] Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.273547 4677 scope.go:117] "RemoveContainer" containerID="d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a" Dec 03 15:50:59 crc kubenswrapper[4677]: E1203 15:50:59.274018 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a\": container with ID starting with d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a not found: ID does not exist" containerID="d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.274077 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a"} err="failed to get container status \"d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a\": rpc error: code = NotFound desc = could not find container \"d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a\": container with ID starting with d8aa3d5532076d41c5ab4edce2fe051063d2a44c2f7a330cd50e95ccbcb5483a not found: ID does not exist" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.274110 4677 scope.go:117] "RemoveContainer" containerID="54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13" Dec 03 15:50:59 crc kubenswrapper[4677]: E1203 15:50:59.274467 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13\": container with ID starting with 54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13 not found: ID does not exist" containerID="54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.274503 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13"} err="failed to get container status \"54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13\": rpc error: code = NotFound desc = could not find container \"54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13\": container with ID starting with 54552dab2e9499d8de21a0233f3512931fec54d4f05bfa75f9c549e900eada13 not found: ID does not exist" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.274523 4677 scope.go:117] "RemoveContainer" containerID="72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4" Dec 03 15:50:59 crc kubenswrapper[4677]: E1203 15:50:59.274730 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4\": container with ID starting with 72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4 not found: ID does not exist" containerID="72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.274758 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4"} err="failed to get container status \"72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4\": rpc error: code = NotFound desc = could not find container \"72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4\": container with ID starting with 72f650c5915d211383f3a70e561deec1e27ecb40d2a336928e5fee74b11a92a4 not found: ID does not exist" Dec 03 15:50:59 crc kubenswrapper[4677]: I1203 15:50:59.995445 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5aa690-9273-42de-a62a-028235952ad2" path="/var/lib/kubelet/pods/ff5aa690-9273-42de-a62a-028235952ad2/volumes" Dec 03 15:51:04 crc kubenswrapper[4677]: I1203 15:51:04.372388 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5rflb_4b745bd9-ab3c-42d4-9476-5460897f8164/nmstate-console-plugin/0.log" Dec 03 15:51:04 crc kubenswrapper[4677]: I1203 15:51:04.593470 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-ghccd_24ca8470-a377-4746-a521-8af29247ef95/kube-rbac-proxy/0.log" Dec 03 15:51:04 crc kubenswrapper[4677]: I1203 15:51:04.630679 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-8bpr7_feb427e2-dfd2-4de1-9f57-e230be83a451/nmstate-handler/0.log" Dec 03 15:51:04 crc kubenswrapper[4677]: I1203 15:51:04.656703 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-ghccd_24ca8470-a377-4746-a521-8af29247ef95/nmstate-metrics/0.log" Dec 03 15:51:04 crc kubenswrapper[4677]: I1203 15:51:04.857311 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-2x69c_a40313ea-273b-4c40-b414-d492a94efb8b/nmstate-operator/0.log" Dec 03 15:51:04 crc kubenswrapper[4677]: I1203 15:51:04.891434 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-zlfwn_80447395-7700-4ff3-9e8c-f18252cf98eb/nmstate-webhook/0.log" Dec 03 15:51:09 crc kubenswrapper[4677]: I1203 15:51:09.988652 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:51:09 crc kubenswrapper[4677]: E1203 15:51:09.990474 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:51:19 crc kubenswrapper[4677]: I1203 15:51:19.975494 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-pb5w7_ae71aaf3-8251-4388-ac3a-26121f63a2c0/kube-rbac-proxy/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.130501 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-pb5w7_ae71aaf3-8251-4388-ac3a-26121f63a2c0/controller/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.234490 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-frr-files/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.405277 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-reloader/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.461223 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-reloader/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.479048 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-metrics/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.491206 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-frr-files/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.636161 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-frr-files/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.662604 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-reloader/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.667995 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-metrics/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.711566 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-metrics/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.924334 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-reloader/0.log" Dec 03 15:51:20 crc kubenswrapper[4677]: I1203 15:51:20.935483 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-metrics/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.180134 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/cp-frr-files/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.352184 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/frr-metrics/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.356548 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/controller/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.421536 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/kube-rbac-proxy/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.540485 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/reloader/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.569801 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/kube-rbac-proxy-frr/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.774623 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-mq57q_b72d11be-cd0b-4d3f-a4d0-068b9e14b4bb/frr-k8s-webhook-server/0.log" Dec 03 15:51:21 crc kubenswrapper[4677]: I1203 15:51:21.893394 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5f48b5ddf5-58d4r_f6eaef50-3b3e-4fb9-a22a-7db928062fa3/manager/1.log" Dec 03 15:51:22 crc kubenswrapper[4677]: I1203 15:51:22.040701 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5f48b5ddf5-58d4r_f6eaef50-3b3e-4fb9-a22a-7db928062fa3/manager/0.log" Dec 03 15:51:22 crc kubenswrapper[4677]: I1203 15:51:22.184626 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5dcf4b8dc4-2f58c_9427b27d-5375-4914-9481-fbbd6c1f9e21/webhook-server/0.log" Dec 03 15:51:22 crc kubenswrapper[4677]: I1203 15:51:22.327521 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4s5lm_630fc3ca-d708-4a40-a6a0-dc2047d01769/kube-rbac-proxy/0.log" Dec 03 15:51:23 crc kubenswrapper[4677]: I1203 15:51:23.048031 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4s5lm_630fc3ca-d708-4a40-a6a0-dc2047d01769/speaker/0.log" Dec 03 15:51:23 crc kubenswrapper[4677]: I1203 15:51:23.541275 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lcgmk_968d4f67-d25c-48f4-b120-a345dc551e55/frr/0.log" Dec 03 15:51:23 crc kubenswrapper[4677]: I1203 15:51:23.976461 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:51:23 crc kubenswrapper[4677]: E1203 15:51:23.976752 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:51:36 crc kubenswrapper[4677]: I1203 15:51:36.976171 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:51:36 crc kubenswrapper[4677]: E1203 15:51:36.977046 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.032825 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x_bbf3592d-dd63-4344-8ae0-2b3b0e88b223/util/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.224049 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x_bbf3592d-dd63-4344-8ae0-2b3b0e88b223/pull/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.225818 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x_bbf3592d-dd63-4344-8ae0-2b3b0e88b223/util/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.272399 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x_bbf3592d-dd63-4344-8ae0-2b3b0e88b223/pull/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.452863 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x_bbf3592d-dd63-4344-8ae0-2b3b0e88b223/extract/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.467977 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x_bbf3592d-dd63-4344-8ae0-2b3b0e88b223/pull/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.491545 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs7d5x_bbf3592d-dd63-4344-8ae0-2b3b0e88b223/util/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.651435 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv_82e2ca22-05b0-4755-a323-1ca6afd42937/util/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.902648 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv_82e2ca22-05b0-4755-a323-1ca6afd42937/util/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.925167 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv_82e2ca22-05b0-4755-a323-1ca6afd42937/pull/0.log" Dec 03 15:51:37 crc kubenswrapper[4677]: I1203 15:51:37.929184 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv_82e2ca22-05b0-4755-a323-1ca6afd42937/pull/0.log" Dec 03 15:51:38 crc kubenswrapper[4677]: I1203 15:51:38.169361 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv_82e2ca22-05b0-4755-a323-1ca6afd42937/util/0.log" Dec 03 15:51:38 crc kubenswrapper[4677]: I1203 15:51:38.216582 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv_82e2ca22-05b0-4755-a323-1ca6afd42937/pull/0.log" Dec 03 15:51:38 crc kubenswrapper[4677]: I1203 15:51:38.227821 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rl5pv_82e2ca22-05b0-4755-a323-1ca6afd42937/extract/0.log" Dec 03 15:51:38 crc kubenswrapper[4677]: I1203 15:51:38.378719 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp_ef1bb5ce-06d7-47db-a43c-85e57f070cff/util/0.log" Dec 03 15:51:38 crc kubenswrapper[4677]: I1203 15:51:38.658195 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp_ef1bb5ce-06d7-47db-a43c-85e57f070cff/util/0.log" Dec 03 15:51:38 crc kubenswrapper[4677]: I1203 15:51:38.884052 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp_ef1bb5ce-06d7-47db-a43c-85e57f070cff/util/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.238851 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp_ef1bb5ce-06d7-47db-a43c-85e57f070cff/pull/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.239071 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp_ef1bb5ce-06d7-47db-a43c-85e57f070cff/pull/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.239122 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp_ef1bb5ce-06d7-47db-a43c-85e57f070cff/pull/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.240195 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83gblzp_ef1bb5ce-06d7-47db-a43c-85e57f070cff/extract/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.490663 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hm4kc_5b895f39-0aeb-4a61-857d-9d9e91b103f3/extract-utilities/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.688986 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hm4kc_5b895f39-0aeb-4a61-857d-9d9e91b103f3/extract-content/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.713153 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hm4kc_5b895f39-0aeb-4a61-857d-9d9e91b103f3/extract-content/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.713226 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hm4kc_5b895f39-0aeb-4a61-857d-9d9e91b103f3/extract-utilities/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.920499 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hm4kc_5b895f39-0aeb-4a61-857d-9d9e91b103f3/extract-content/0.log" Dec 03 15:51:39 crc kubenswrapper[4677]: I1203 15:51:39.925060 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hm4kc_5b895f39-0aeb-4a61-857d-9d9e91b103f3/extract-utilities/0.log" Dec 03 15:51:40 crc kubenswrapper[4677]: I1203 15:51:40.160042 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-68vg8_361c4f16-5752-4a54-bae4-033bc1425b51/extract-utilities/0.log" Dec 03 15:51:40 crc kubenswrapper[4677]: I1203 15:51:40.398525 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-68vg8_361c4f16-5752-4a54-bae4-033bc1425b51/extract-utilities/0.log" Dec 03 15:51:40 crc kubenswrapper[4677]: I1203 15:51:40.521258 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-68vg8_361c4f16-5752-4a54-bae4-033bc1425b51/extract-content/0.log" Dec 03 15:51:40 crc kubenswrapper[4677]: I1203 15:51:40.537921 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-68vg8_361c4f16-5752-4a54-bae4-033bc1425b51/extract-content/0.log" Dec 03 15:51:40 crc kubenswrapper[4677]: I1203 15:51:40.723256 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-68vg8_361c4f16-5752-4a54-bae4-033bc1425b51/extract-content/0.log" Dec 03 15:51:40 crc kubenswrapper[4677]: I1203 15:51:40.724935 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-68vg8_361c4f16-5752-4a54-bae4-033bc1425b51/extract-utilities/0.log" Dec 03 15:51:41 crc kubenswrapper[4677]: I1203 15:51:41.011213 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s4dpj_d1c06aff-5cb1-4653-a4ed-bfddfab19938/extract-utilities/0.log" Dec 03 15:51:41 crc kubenswrapper[4677]: I1203 15:51:41.125788 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tfhnk_ed0d9b44-1fd5-4c95-b064-5bd1fad3e3f6/marketplace-operator/0.log" Dec 03 15:51:41 crc kubenswrapper[4677]: I1203 15:51:41.131339 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s4dpj_d1c06aff-5cb1-4653-a4ed-bfddfab19938/extract-utilities/0.log" Dec 03 15:51:41 crc kubenswrapper[4677]: I1203 15:51:41.174441 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s4dpj_d1c06aff-5cb1-4653-a4ed-bfddfab19938/extract-content/0.log" Dec 03 15:51:41 crc kubenswrapper[4677]: I1203 15:51:41.510727 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s4dpj_d1c06aff-5cb1-4653-a4ed-bfddfab19938/extract-content/0.log" Dec 03 15:51:41 crc kubenswrapper[4677]: I1203 15:51:41.893518 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s4dpj_d1c06aff-5cb1-4653-a4ed-bfddfab19938/extract-content/0.log" Dec 03 15:51:41 crc kubenswrapper[4677]: I1203 15:51:41.967108 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s4dpj_d1c06aff-5cb1-4653-a4ed-bfddfab19938/extract-utilities/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.050964 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hm4kc_5b895f39-0aeb-4a61-857d-9d9e91b103f3/registry-server/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.171897 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-785mp_68ad29ee-1d1c-41ac-8679-9b30bd41e1a6/extract-utilities/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.397641 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-785mp_68ad29ee-1d1c-41ac-8679-9b30bd41e1a6/extract-utilities/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.548046 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-785mp_68ad29ee-1d1c-41ac-8679-9b30bd41e1a6/extract-content/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.579212 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-785mp_68ad29ee-1d1c-41ac-8679-9b30bd41e1a6/extract-content/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.608108 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-s4dpj_d1c06aff-5cb1-4653-a4ed-bfddfab19938/registry-server/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.790513 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-785mp_68ad29ee-1d1c-41ac-8679-9b30bd41e1a6/extract-content/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.814246 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-785mp_68ad29ee-1d1c-41ac-8679-9b30bd41e1a6/extract-utilities/0.log" Dec 03 15:51:42 crc kubenswrapper[4677]: I1203 15:51:42.961320 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-68vg8_361c4f16-5752-4a54-bae4-033bc1425b51/registry-server/0.log" Dec 03 15:51:43 crc kubenswrapper[4677]: I1203 15:51:43.698448 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-785mp_68ad29ee-1d1c-41ac-8679-9b30bd41e1a6/registry-server/0.log" Dec 03 15:51:51 crc kubenswrapper[4677]: I1203 15:51:51.996716 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:51:51 crc kubenswrapper[4677]: E1203 15:51:51.997512 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:51:57 crc kubenswrapper[4677]: I1203 15:51:57.840875 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-pkpj9_61643dea-c510-4111-8ab1-ec1afc438078/prometheus-operator/0.log" Dec 03 15:51:58 crc kubenswrapper[4677]: I1203 15:51:58.060211 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67cbb44f74-f6mlm_b737ec59-d50e-4179-ac09-1bfa1e51529c/prometheus-operator-admission-webhook/0.log" Dec 03 15:51:58 crc kubenswrapper[4677]: I1203 15:51:58.073385 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67cbb44f74-ljd4g_16fd0377-c577-4f4e-8e02-5d1e31718895/prometheus-operator-admission-webhook/0.log" Dec 03 15:51:58 crc kubenswrapper[4677]: I1203 15:51:58.247602 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-ps6w5_65806097-c974-427b-a63c-ed84eaa3b039/operator/0.log" Dec 03 15:51:58 crc kubenswrapper[4677]: I1203 15:51:58.337320 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-lm89l_cfed8170-869b-48c2-8b7f-e69466674721/perses-operator/0.log" Dec 03 15:52:06 crc kubenswrapper[4677]: I1203 15:52:06.975773 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:52:06 crc kubenswrapper[4677]: E1203 15:52:06.976407 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:52:17 crc kubenswrapper[4677]: I1203 15:52:17.977646 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:52:17 crc kubenswrapper[4677]: E1203 15:52:17.978264 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:52:25 crc kubenswrapper[4677]: E1203 15:52:25.570050 4677 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.162:55422->38.102.83.162:39005: write tcp 38.102.83.162:55422->38.102.83.162:39005: write: broken pipe Dec 03 15:52:30 crc kubenswrapper[4677]: I1203 15:52:30.975918 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:52:30 crc kubenswrapper[4677]: E1203 15:52:30.976923 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:52:42 crc kubenswrapper[4677]: I1203 15:52:42.977270 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:52:42 crc kubenswrapper[4677]: E1203 15:52:42.977987 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:52:57 crc kubenswrapper[4677]: I1203 15:52:57.976773 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:52:57 crc kubenswrapper[4677]: E1203 15:52:57.977518 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:53:08 crc kubenswrapper[4677]: I1203 15:53:08.977021 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:53:08 crc kubenswrapper[4677]: E1203 15:53:08.977768 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:53:19 crc kubenswrapper[4677]: I1203 15:53:19.987854 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:53:19 crc kubenswrapper[4677]: E1203 15:53:19.990019 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:53:32 crc kubenswrapper[4677]: I1203 15:53:32.976737 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:53:32 crc kubenswrapper[4677]: E1203 15:53:32.977575 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 15:53:45 crc kubenswrapper[4677]: I1203 15:53:45.977677 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:53:47 crc kubenswrapper[4677]: I1203 15:53:47.006919 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"4bf1d337bf6e33dc878f09d0f5a4457a9b07f765334ef5c2d68bc844e60b6241"} Dec 03 15:54:03 crc kubenswrapper[4677]: I1203 15:54:03.164967 4677 generic.go:334] "Generic (PLEG): container finished" podID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerID="f7adac2a771fcee555002cc5232e310021957c82f9815308ed6eb3f49bf74c47" exitCode=0 Dec 03 15:54:03 crc kubenswrapper[4677]: I1203 15:54:03.165040 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rfmhp/must-gather-zh684" event={"ID":"34e26d55-0f57-4199-8a80-b9a009649d8f","Type":"ContainerDied","Data":"f7adac2a771fcee555002cc5232e310021957c82f9815308ed6eb3f49bf74c47"} Dec 03 15:54:03 crc kubenswrapper[4677]: I1203 15:54:03.166314 4677 scope.go:117] "RemoveContainer" containerID="f7adac2a771fcee555002cc5232e310021957c82f9815308ed6eb3f49bf74c47" Dec 03 15:54:03 crc kubenswrapper[4677]: I1203 15:54:03.922909 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rfmhp_must-gather-zh684_34e26d55-0f57-4199-8a80-b9a009649d8f/gather/0.log" Dec 03 15:54:12 crc kubenswrapper[4677]: I1203 15:54:12.935978 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rfmhp/must-gather-zh684"] Dec 03 15:54:12 crc kubenswrapper[4677]: I1203 15:54:12.936666 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rfmhp/must-gather-zh684" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerName="copy" containerID="cri-o://2598f87f9be583e35577869a9c7f1929190ce651ab7139ce90b6b739eb7ac519" gracePeriod=2 Dec 03 15:54:12 crc kubenswrapper[4677]: I1203 15:54:12.948351 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rfmhp/must-gather-zh684"] Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.290289 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rfmhp_must-gather-zh684_34e26d55-0f57-4199-8a80-b9a009649d8f/copy/0.log" Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.291045 4677 generic.go:334] "Generic (PLEG): container finished" podID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerID="2598f87f9be583e35577869a9c7f1929190ce651ab7139ce90b6b739eb7ac519" exitCode=143 Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.526717 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rfmhp_must-gather-zh684_34e26d55-0f57-4199-8a80-b9a009649d8f/copy/0.log" Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.527223 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.691536 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34e26d55-0f57-4199-8a80-b9a009649d8f-must-gather-output\") pod \"34e26d55-0f57-4199-8a80-b9a009649d8f\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.692247 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcmzl\" (UniqueName: \"kubernetes.io/projected/34e26d55-0f57-4199-8a80-b9a009649d8f-kube-api-access-mcmzl\") pod \"34e26d55-0f57-4199-8a80-b9a009649d8f\" (UID: \"34e26d55-0f57-4199-8a80-b9a009649d8f\") " Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.699749 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34e26d55-0f57-4199-8a80-b9a009649d8f-kube-api-access-mcmzl" (OuterVolumeSpecName: "kube-api-access-mcmzl") pod "34e26d55-0f57-4199-8a80-b9a009649d8f" (UID: "34e26d55-0f57-4199-8a80-b9a009649d8f"). InnerVolumeSpecName "kube-api-access-mcmzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.795286 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcmzl\" (UniqueName: \"kubernetes.io/projected/34e26d55-0f57-4199-8a80-b9a009649d8f-kube-api-access-mcmzl\") on node \"crc\" DevicePath \"\"" Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.905046 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34e26d55-0f57-4199-8a80-b9a009649d8f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "34e26d55-0f57-4199-8a80-b9a009649d8f" (UID: "34e26d55-0f57-4199-8a80-b9a009649d8f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:54:13 crc kubenswrapper[4677]: I1203 15:54:13.994703 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" path="/var/lib/kubelet/pods/34e26d55-0f57-4199-8a80-b9a009649d8f/volumes" Dec 03 15:54:14 crc kubenswrapper[4677]: I1203 15:54:14.004438 4677 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/34e26d55-0f57-4199-8a80-b9a009649d8f-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 15:54:14 crc kubenswrapper[4677]: I1203 15:54:14.318184 4677 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rfmhp_must-gather-zh684_34e26d55-0f57-4199-8a80-b9a009649d8f/copy/0.log" Dec 03 15:54:14 crc kubenswrapper[4677]: I1203 15:54:14.318629 4677 scope.go:117] "RemoveContainer" containerID="2598f87f9be583e35577869a9c7f1929190ce651ab7139ce90b6b739eb7ac519" Dec 03 15:54:14 crc kubenswrapper[4677]: I1203 15:54:14.318759 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rfmhp/must-gather-zh684" Dec 03 15:54:14 crc kubenswrapper[4677]: I1203 15:54:14.361197 4677 scope.go:117] "RemoveContainer" containerID="f7adac2a771fcee555002cc5232e310021957c82f9815308ed6eb3f49bf74c47" Dec 03 15:55:53 crc kubenswrapper[4677]: I1203 15:55:53.114608 4677 scope.go:117] "RemoveContainer" containerID="6de582ee9edefa7393a23e88815eec4c091afc4f54c5179b670507ccdf23fba7" Dec 03 15:56:08 crc kubenswrapper[4677]: I1203 15:56:08.437340 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:56:08 crc kubenswrapper[4677]: I1203 15:56:08.437994 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.351425 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4b5t"] Dec 03 15:56:19 crc kubenswrapper[4677]: E1203 15:56:19.352430 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="extract-content" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352443 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="extract-content" Dec 03 15:56:19 crc kubenswrapper[4677]: E1203 15:56:19.352459 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerName="copy" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352465 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerName="copy" Dec 03 15:56:19 crc kubenswrapper[4677]: E1203 15:56:19.352481 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerName="gather" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352486 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerName="gather" Dec 03 15:56:19 crc kubenswrapper[4677]: E1203 15:56:19.352502 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="extract-utilities" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352510 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="extract-utilities" Dec 03 15:56:19 crc kubenswrapper[4677]: E1203 15:56:19.352529 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="registry-server" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352537 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="registry-server" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352756 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5aa690-9273-42de-a62a-028235952ad2" containerName="registry-server" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352773 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerName="copy" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.352787 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e26d55-0f57-4199-8a80-b9a009649d8f" containerName="gather" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.354584 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.368305 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4b5t"] Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.489347 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-catalog-content\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.489445 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6z28\" (UniqueName: \"kubernetes.io/projected/a3a2f2d2-84d8-459f-9c29-dc3e92607453-kube-api-access-d6z28\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.489541 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-utilities\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.591991 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-utilities\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.592189 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-catalog-content\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.592237 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6z28\" (UniqueName: \"kubernetes.io/projected/a3a2f2d2-84d8-459f-9c29-dc3e92607453-kube-api-access-d6z28\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.592898 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-utilities\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.593051 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-catalog-content\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.622844 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6z28\" (UniqueName: \"kubernetes.io/projected/a3a2f2d2-84d8-459f-9c29-dc3e92607453-kube-api-access-d6z28\") pod \"community-operators-p4b5t\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:19 crc kubenswrapper[4677]: I1203 15:56:19.691299 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:20 crc kubenswrapper[4677]: I1203 15:56:20.230092 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4b5t"] Dec 03 15:56:21 crc kubenswrapper[4677]: I1203 15:56:21.033479 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerID="d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd" exitCode=0 Dec 03 15:56:21 crc kubenswrapper[4677]: I1203 15:56:21.033522 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4b5t" event={"ID":"a3a2f2d2-84d8-459f-9c29-dc3e92607453","Type":"ContainerDied","Data":"d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd"} Dec 03 15:56:21 crc kubenswrapper[4677]: I1203 15:56:21.033548 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4b5t" event={"ID":"a3a2f2d2-84d8-459f-9c29-dc3e92607453","Type":"ContainerStarted","Data":"6206b07128d20934234077befcd01c1da35fda8324385b9b7fd54eb0531b4d94"} Dec 03 15:56:21 crc kubenswrapper[4677]: I1203 15:56:21.035611 4677 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:56:22 crc kubenswrapper[4677]: I1203 15:56:22.049423 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4b5t" event={"ID":"a3a2f2d2-84d8-459f-9c29-dc3e92607453","Type":"ContainerStarted","Data":"f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959"} Dec 03 15:56:23 crc kubenswrapper[4677]: I1203 15:56:23.060167 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerID="f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959" exitCode=0 Dec 03 15:56:23 crc kubenswrapper[4677]: I1203 15:56:23.060221 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4b5t" event={"ID":"a3a2f2d2-84d8-459f-9c29-dc3e92607453","Type":"ContainerDied","Data":"f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959"} Dec 03 15:56:24 crc kubenswrapper[4677]: I1203 15:56:24.074414 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4b5t" event={"ID":"a3a2f2d2-84d8-459f-9c29-dc3e92607453","Type":"ContainerStarted","Data":"c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17"} Dec 03 15:56:24 crc kubenswrapper[4677]: I1203 15:56:24.107378 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4b5t" podStartSLOduration=2.622995716 podStartE2EDuration="5.107354627s" podCreationTimestamp="2025-12-03 15:56:19 +0000 UTC" firstStartedPulling="2025-12-03 15:56:21.035216914 +0000 UTC m=+7771.781549369" lastFinishedPulling="2025-12-03 15:56:23.519575835 +0000 UTC m=+7774.265908280" observedRunningTime="2025-12-03 15:56:24.096566298 +0000 UTC m=+7774.842898793" watchObservedRunningTime="2025-12-03 15:56:24.107354627 +0000 UTC m=+7774.853687122" Dec 03 15:56:29 crc kubenswrapper[4677]: I1203 15:56:29.691667 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:29 crc kubenswrapper[4677]: I1203 15:56:29.692259 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:29 crc kubenswrapper[4677]: I1203 15:56:29.742183 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:30 crc kubenswrapper[4677]: I1203 15:56:30.203371 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:30 crc kubenswrapper[4677]: I1203 15:56:30.256967 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4b5t"] Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.173231 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p4b5t" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="registry-server" containerID="cri-o://c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17" gracePeriod=2 Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.747357 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.890769 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-catalog-content\") pod \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.890828 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6z28\" (UniqueName: \"kubernetes.io/projected/a3a2f2d2-84d8-459f-9c29-dc3e92607453-kube-api-access-d6z28\") pod \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.891011 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-utilities\") pod \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\" (UID: \"a3a2f2d2-84d8-459f-9c29-dc3e92607453\") " Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.891792 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-utilities" (OuterVolumeSpecName: "utilities") pod "a3a2f2d2-84d8-459f-9c29-dc3e92607453" (UID: "a3a2f2d2-84d8-459f-9c29-dc3e92607453"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.897070 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a2f2d2-84d8-459f-9c29-dc3e92607453-kube-api-access-d6z28" (OuterVolumeSpecName: "kube-api-access-d6z28") pod "a3a2f2d2-84d8-459f-9c29-dc3e92607453" (UID: "a3a2f2d2-84d8-459f-9c29-dc3e92607453"). InnerVolumeSpecName "kube-api-access-d6z28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.943894 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3a2f2d2-84d8-459f-9c29-dc3e92607453" (UID: "a3a2f2d2-84d8-459f-9c29-dc3e92607453"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.992972 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.993004 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6z28\" (UniqueName: \"kubernetes.io/projected/a3a2f2d2-84d8-459f-9c29-dc3e92607453-kube-api-access-d6z28\") on node \"crc\" DevicePath \"\"" Dec 03 15:56:32 crc kubenswrapper[4677]: I1203 15:56:32.993014 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a2f2d2-84d8-459f-9c29-dc3e92607453-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.185696 4677 generic.go:334] "Generic (PLEG): container finished" podID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerID="c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17" exitCode=0 Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.185765 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4b5t" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.185756 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4b5t" event={"ID":"a3a2f2d2-84d8-459f-9c29-dc3e92607453","Type":"ContainerDied","Data":"c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17"} Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.186169 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4b5t" event={"ID":"a3a2f2d2-84d8-459f-9c29-dc3e92607453","Type":"ContainerDied","Data":"6206b07128d20934234077befcd01c1da35fda8324385b9b7fd54eb0531b4d94"} Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.186199 4677 scope.go:117] "RemoveContainer" containerID="c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.219280 4677 scope.go:117] "RemoveContainer" containerID="f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.238066 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4b5t"] Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.250750 4677 scope.go:117] "RemoveContainer" containerID="d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.252681 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p4b5t"] Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.310804 4677 scope.go:117] "RemoveContainer" containerID="c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17" Dec 03 15:56:33 crc kubenswrapper[4677]: E1203 15:56:33.311415 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17\": container with ID starting with c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17 not found: ID does not exist" containerID="c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.311469 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17"} err="failed to get container status \"c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17\": rpc error: code = NotFound desc = could not find container \"c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17\": container with ID starting with c33ace3ec21194aa4b3887c29779ee035eaddaa90207f929cb3e48900bce1d17 not found: ID does not exist" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.311504 4677 scope.go:117] "RemoveContainer" containerID="f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959" Dec 03 15:56:33 crc kubenswrapper[4677]: E1203 15:56:33.311992 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959\": container with ID starting with f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959 not found: ID does not exist" containerID="f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.312017 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959"} err="failed to get container status \"f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959\": rpc error: code = NotFound desc = could not find container \"f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959\": container with ID starting with f5958a482e5f5f45171cf553a32c575e32abc042a8bedf3e12ec598123dce959 not found: ID does not exist" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.312037 4677 scope.go:117] "RemoveContainer" containerID="d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd" Dec 03 15:56:33 crc kubenswrapper[4677]: E1203 15:56:33.312328 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd\": container with ID starting with d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd not found: ID does not exist" containerID="d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.312354 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd"} err="failed to get container status \"d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd\": rpc error: code = NotFound desc = could not find container \"d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd\": container with ID starting with d105662d2785737a89d7f81251197ed8e1c3f05c9b4486bb23ed60bf2966d3fd not found: ID does not exist" Dec 03 15:56:33 crc kubenswrapper[4677]: I1203 15:56:33.989295 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" path="/var/lib/kubelet/pods/a3a2f2d2-84d8-459f-9c29-dc3e92607453/volumes" Dec 03 15:56:38 crc kubenswrapper[4677]: I1203 15:56:38.437833 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:56:38 crc kubenswrapper[4677]: I1203 15:56:38.438309 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.437932 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.439337 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.439415 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.440897 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4bf1d337bf6e33dc878f09d0f5a4457a9b07f765334ef5c2d68bc844e60b6241"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.441019 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://4bf1d337bf6e33dc878f09d0f5a4457a9b07f765334ef5c2d68bc844e60b6241" gracePeriod=600 Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.567177 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="4bf1d337bf6e33dc878f09d0f5a4457a9b07f765334ef5c2d68bc844e60b6241" exitCode=0 Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.567260 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"4bf1d337bf6e33dc878f09d0f5a4457a9b07f765334ef5c2d68bc844e60b6241"} Dec 03 15:57:08 crc kubenswrapper[4677]: I1203 15:57:08.567645 4677 scope.go:117] "RemoveContainer" containerID="1505ca5009abe4c1ddf2f29788cc78b4b43652346b3124cb6d7e5ef269b84f6b" Dec 03 15:57:09 crc kubenswrapper[4677]: I1203 15:57:09.578470 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerStarted","Data":"573e5375928362f9817cbfab82c4bb837d24528e023e4f9d74d58886defab8fa"} Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.679922 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p8fdf"] Dec 03 15:57:31 crc kubenswrapper[4677]: E1203 15:57:31.681383 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="extract-utilities" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.681414 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="extract-utilities" Dec 03 15:57:31 crc kubenswrapper[4677]: E1203 15:57:31.681436 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="extract-content" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.681447 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="extract-content" Dec 03 15:57:31 crc kubenswrapper[4677]: E1203 15:57:31.681497 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="registry-server" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.681506 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="registry-server" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.682342 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a2f2d2-84d8-459f-9c29-dc3e92607453" containerName="registry-server" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.686402 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.697765 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p8fdf"] Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.862793 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-catalog-content\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.862914 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nmnq\" (UniqueName: \"kubernetes.io/projected/473d65ed-45c5-4949-b751-43f59d22939b-kube-api-access-8nmnq\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.862938 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-utilities\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.965233 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-catalog-content\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.965364 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nmnq\" (UniqueName: \"kubernetes.io/projected/473d65ed-45c5-4949-b751-43f59d22939b-kube-api-access-8nmnq\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.965387 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-utilities\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.965992 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-catalog-content\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.966024 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-utilities\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:31 crc kubenswrapper[4677]: I1203 15:57:31.991756 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nmnq\" (UniqueName: \"kubernetes.io/projected/473d65ed-45c5-4949-b751-43f59d22939b-kube-api-access-8nmnq\") pod \"redhat-operators-p8fdf\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:32 crc kubenswrapper[4677]: I1203 15:57:32.013839 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:32 crc kubenswrapper[4677]: I1203 15:57:32.555525 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p8fdf"] Dec 03 15:57:32 crc kubenswrapper[4677]: I1203 15:57:32.804032 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8fdf" event={"ID":"473d65ed-45c5-4949-b751-43f59d22939b","Type":"ContainerStarted","Data":"1b5b5fad01506d57b6bb6528bcf0d7e95a6ac48d3257e9e8f14d20c0c8fcaff9"} Dec 03 15:57:33 crc kubenswrapper[4677]: I1203 15:57:33.814893 4677 generic.go:334] "Generic (PLEG): container finished" podID="473d65ed-45c5-4949-b751-43f59d22939b" containerID="e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0" exitCode=0 Dec 03 15:57:33 crc kubenswrapper[4677]: I1203 15:57:33.815016 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8fdf" event={"ID":"473d65ed-45c5-4949-b751-43f59d22939b","Type":"ContainerDied","Data":"e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0"} Dec 03 15:57:34 crc kubenswrapper[4677]: I1203 15:57:34.834214 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8fdf" event={"ID":"473d65ed-45c5-4949-b751-43f59d22939b","Type":"ContainerStarted","Data":"83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68"} Dec 03 15:57:38 crc kubenswrapper[4677]: I1203 15:57:38.879327 4677 generic.go:334] "Generic (PLEG): container finished" podID="473d65ed-45c5-4949-b751-43f59d22939b" containerID="83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68" exitCode=0 Dec 03 15:57:38 crc kubenswrapper[4677]: I1203 15:57:38.879421 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8fdf" event={"ID":"473d65ed-45c5-4949-b751-43f59d22939b","Type":"ContainerDied","Data":"83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68"} Dec 03 15:57:39 crc kubenswrapper[4677]: I1203 15:57:39.896173 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8fdf" event={"ID":"473d65ed-45c5-4949-b751-43f59d22939b","Type":"ContainerStarted","Data":"328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937"} Dec 03 15:57:39 crc kubenswrapper[4677]: I1203 15:57:39.926686 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p8fdf" podStartSLOduration=3.462219314 podStartE2EDuration="8.926663553s" podCreationTimestamp="2025-12-03 15:57:31 +0000 UTC" firstStartedPulling="2025-12-03 15:57:33.817269776 +0000 UTC m=+7844.563602231" lastFinishedPulling="2025-12-03 15:57:39.281714015 +0000 UTC m=+7850.028046470" observedRunningTime="2025-12-03 15:57:39.916975044 +0000 UTC m=+7850.663307509" watchObservedRunningTime="2025-12-03 15:57:39.926663553 +0000 UTC m=+7850.672996008" Dec 03 15:57:42 crc kubenswrapper[4677]: I1203 15:57:42.014829 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:42 crc kubenswrapper[4677]: I1203 15:57:42.016288 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:43 crc kubenswrapper[4677]: I1203 15:57:43.062982 4677 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p8fdf" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="registry-server" probeResult="failure" output=< Dec 03 15:57:43 crc kubenswrapper[4677]: timeout: failed to connect service ":50051" within 1s Dec 03 15:57:43 crc kubenswrapper[4677]: > Dec 03 15:57:52 crc kubenswrapper[4677]: I1203 15:57:52.131219 4677 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:52 crc kubenswrapper[4677]: I1203 15:57:52.197659 4677 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:52 crc kubenswrapper[4677]: I1203 15:57:52.371033 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p8fdf"] Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.059218 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p8fdf" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="registry-server" containerID="cri-o://328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937" gracePeriod=2 Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.660244 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.792675 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-utilities\") pod \"473d65ed-45c5-4949-b751-43f59d22939b\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.792846 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nmnq\" (UniqueName: \"kubernetes.io/projected/473d65ed-45c5-4949-b751-43f59d22939b-kube-api-access-8nmnq\") pod \"473d65ed-45c5-4949-b751-43f59d22939b\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.792972 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-catalog-content\") pod \"473d65ed-45c5-4949-b751-43f59d22939b\" (UID: \"473d65ed-45c5-4949-b751-43f59d22939b\") " Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.793795 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-utilities" (OuterVolumeSpecName: "utilities") pod "473d65ed-45c5-4949-b751-43f59d22939b" (UID: "473d65ed-45c5-4949-b751-43f59d22939b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.799663 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/473d65ed-45c5-4949-b751-43f59d22939b-kube-api-access-8nmnq" (OuterVolumeSpecName: "kube-api-access-8nmnq") pod "473d65ed-45c5-4949-b751-43f59d22939b" (UID: "473d65ed-45c5-4949-b751-43f59d22939b"). InnerVolumeSpecName "kube-api-access-8nmnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.895236 4677 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.895268 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nmnq\" (UniqueName: \"kubernetes.io/projected/473d65ed-45c5-4949-b751-43f59d22939b-kube-api-access-8nmnq\") on node \"crc\" DevicePath \"\"" Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.900318 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "473d65ed-45c5-4949-b751-43f59d22939b" (UID: "473d65ed-45c5-4949-b751-43f59d22939b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:57:54 crc kubenswrapper[4677]: I1203 15:57:54.996771 4677 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/473d65ed-45c5-4949-b751-43f59d22939b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.068622 4677 generic.go:334] "Generic (PLEG): container finished" podID="473d65ed-45c5-4949-b751-43f59d22939b" containerID="328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937" exitCode=0 Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.068667 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8fdf" event={"ID":"473d65ed-45c5-4949-b751-43f59d22939b","Type":"ContainerDied","Data":"328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937"} Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.068694 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8fdf" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.068715 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8fdf" event={"ID":"473d65ed-45c5-4949-b751-43f59d22939b","Type":"ContainerDied","Data":"1b5b5fad01506d57b6bb6528bcf0d7e95a6ac48d3257e9e8f14d20c0c8fcaff9"} Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.068735 4677 scope.go:117] "RemoveContainer" containerID="328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.097253 4677 scope.go:117] "RemoveContainer" containerID="83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.105144 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p8fdf"] Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.121081 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p8fdf"] Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.138449 4677 scope.go:117] "RemoveContainer" containerID="e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.178082 4677 scope.go:117] "RemoveContainer" containerID="328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937" Dec 03 15:57:55 crc kubenswrapper[4677]: E1203 15:57:55.178524 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937\": container with ID starting with 328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937 not found: ID does not exist" containerID="328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.178604 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937"} err="failed to get container status \"328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937\": rpc error: code = NotFound desc = could not find container \"328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937\": container with ID starting with 328928c079aa2c6819edb05f92b7b22dd826b802477a949b76de983086bfa937 not found: ID does not exist" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.178631 4677 scope.go:117] "RemoveContainer" containerID="83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68" Dec 03 15:57:55 crc kubenswrapper[4677]: E1203 15:57:55.179373 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68\": container with ID starting with 83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68 not found: ID does not exist" containerID="83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.179412 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68"} err="failed to get container status \"83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68\": rpc error: code = NotFound desc = could not find container \"83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68\": container with ID starting with 83721b4b0a2c5b5fc797f1ea33d2cbffd62888f1c14ed5e8db52d2d1799abf68 not found: ID does not exist" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.179439 4677 scope.go:117] "RemoveContainer" containerID="e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0" Dec 03 15:57:55 crc kubenswrapper[4677]: E1203 15:57:55.179647 4677 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0\": container with ID starting with e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0 not found: ID does not exist" containerID="e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.179673 4677 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0"} err="failed to get container status \"e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0\": rpc error: code = NotFound desc = could not find container \"e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0\": container with ID starting with e13900824e4ba2a509eaf936e28439837fa33aacebacfd5731ca96605af5dfa0 not found: ID does not exist" Dec 03 15:57:55 crc kubenswrapper[4677]: I1203 15:57:55.988350 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="473d65ed-45c5-4949-b751-43f59d22939b" path="/var/lib/kubelet/pods/473d65ed-45c5-4949-b751-43f59d22939b/volumes" Dec 03 15:59:08 crc kubenswrapper[4677]: I1203 15:59:08.437358 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:59:08 crc kubenswrapper[4677]: I1203 15:59:08.437905 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:59:38 crc kubenswrapper[4677]: I1203 15:59:38.437756 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:59:38 crc kubenswrapper[4677]: I1203 15:59:38.438892 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.159631 4677 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5"] Dec 03 16:00:00 crc kubenswrapper[4677]: E1203 16:00:00.164286 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="registry-server" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.164475 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="registry-server" Dec 03 16:00:00 crc kubenswrapper[4677]: E1203 16:00:00.164529 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="extract-utilities" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.164541 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="extract-utilities" Dec 03 16:00:00 crc kubenswrapper[4677]: E1203 16:00:00.164563 4677 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="extract-content" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.164571 4677 state_mem.go:107] "Deleted CPUSet assignment" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="extract-content" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.165082 4677 memory_manager.go:354] "RemoveStaleState removing state" podUID="473d65ed-45c5-4949-b751-43f59d22939b" containerName="registry-server" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.166460 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.169147 4677 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.169158 4677 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.177387 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5"] Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.199005 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2732e5af-9710-4563-a3c8-056e8eb3941e-config-volume\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.199170 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj66v\" (UniqueName: \"kubernetes.io/projected/2732e5af-9710-4563-a3c8-056e8eb3941e-kube-api-access-wj66v\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.199251 4677 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2732e5af-9710-4563-a3c8-056e8eb3941e-secret-volume\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.300988 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2732e5af-9710-4563-a3c8-056e8eb3941e-secret-volume\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.301194 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2732e5af-9710-4563-a3c8-056e8eb3941e-config-volume\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.301357 4677 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj66v\" (UniqueName: \"kubernetes.io/projected/2732e5af-9710-4563-a3c8-056e8eb3941e-kube-api-access-wj66v\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.302291 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2732e5af-9710-4563-a3c8-056e8eb3941e-config-volume\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.307483 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2732e5af-9710-4563-a3c8-056e8eb3941e-secret-volume\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.324635 4677 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj66v\" (UniqueName: \"kubernetes.io/projected/2732e5af-9710-4563-a3c8-056e8eb3941e-kube-api-access-wj66v\") pod \"collect-profiles-29412960-f5cx5\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.499445 4677 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:00 crc kubenswrapper[4677]: I1203 16:00:00.994428 4677 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5"] Dec 03 16:00:01 crc kubenswrapper[4677]: I1203 16:00:01.445174 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" event={"ID":"2732e5af-9710-4563-a3c8-056e8eb3941e","Type":"ContainerStarted","Data":"a9e429bdb06fd731be0eb5df7a87da327f8e17404640f08b82f943bf71401533"} Dec 03 16:00:01 crc kubenswrapper[4677]: I1203 16:00:01.445533 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" event={"ID":"2732e5af-9710-4563-a3c8-056e8eb3941e","Type":"ContainerStarted","Data":"e022b16bf51f1e5059fe0261194348570e6ab0ade53b09ae7a127c388b5a7dc5"} Dec 03 16:00:01 crc kubenswrapper[4677]: I1203 16:00:01.471119 4677 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" podStartSLOduration=1.471094604 podStartE2EDuration="1.471094604s" podCreationTimestamp="2025-12-03 16:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:00:01.466777316 +0000 UTC m=+7992.213109791" watchObservedRunningTime="2025-12-03 16:00:01.471094604 +0000 UTC m=+7992.217427059" Dec 03 16:00:02 crc kubenswrapper[4677]: I1203 16:00:02.457141 4677 generic.go:334] "Generic (PLEG): container finished" podID="2732e5af-9710-4563-a3c8-056e8eb3941e" containerID="a9e429bdb06fd731be0eb5df7a87da327f8e17404640f08b82f943bf71401533" exitCode=0 Dec 03 16:00:02 crc kubenswrapper[4677]: I1203 16:00:02.457306 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" event={"ID":"2732e5af-9710-4563-a3c8-056e8eb3941e","Type":"ContainerDied","Data":"a9e429bdb06fd731be0eb5df7a87da327f8e17404640f08b82f943bf71401533"} Dec 03 16:00:03 crc kubenswrapper[4677]: I1203 16:00:03.845021 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:03 crc kubenswrapper[4677]: I1203 16:00:03.991509 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj66v\" (UniqueName: \"kubernetes.io/projected/2732e5af-9710-4563-a3c8-056e8eb3941e-kube-api-access-wj66v\") pod \"2732e5af-9710-4563-a3c8-056e8eb3941e\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " Dec 03 16:00:03 crc kubenswrapper[4677]: I1203 16:00:03.991711 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2732e5af-9710-4563-a3c8-056e8eb3941e-secret-volume\") pod \"2732e5af-9710-4563-a3c8-056e8eb3941e\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " Dec 03 16:00:03 crc kubenswrapper[4677]: I1203 16:00:03.991757 4677 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2732e5af-9710-4563-a3c8-056e8eb3941e-config-volume\") pod \"2732e5af-9710-4563-a3c8-056e8eb3941e\" (UID: \"2732e5af-9710-4563-a3c8-056e8eb3941e\") " Dec 03 16:00:03 crc kubenswrapper[4677]: I1203 16:00:03.992680 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2732e5af-9710-4563-a3c8-056e8eb3941e-config-volume" (OuterVolumeSpecName: "config-volume") pod "2732e5af-9710-4563-a3c8-056e8eb3941e" (UID: "2732e5af-9710-4563-a3c8-056e8eb3941e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:00:03 crc kubenswrapper[4677]: I1203 16:00:03.999136 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2732e5af-9710-4563-a3c8-056e8eb3941e-kube-api-access-wj66v" (OuterVolumeSpecName: "kube-api-access-wj66v") pod "2732e5af-9710-4563-a3c8-056e8eb3941e" (UID: "2732e5af-9710-4563-a3c8-056e8eb3941e"). InnerVolumeSpecName "kube-api-access-wj66v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:00:03 crc kubenswrapper[4677]: I1203 16:00:03.999497 4677 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2732e5af-9710-4563-a3c8-056e8eb3941e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2732e5af-9710-4563-a3c8-056e8eb3941e" (UID: "2732e5af-9710-4563-a3c8-056e8eb3941e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.093827 4677 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj66v\" (UniqueName: \"kubernetes.io/projected/2732e5af-9710-4563-a3c8-056e8eb3941e-kube-api-access-wj66v\") on node \"crc\" DevicePath \"\"" Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.093867 4677 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2732e5af-9710-4563-a3c8-056e8eb3941e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.093880 4677 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2732e5af-9710-4563-a3c8-056e8eb3941e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.478235 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" event={"ID":"2732e5af-9710-4563-a3c8-056e8eb3941e","Type":"ContainerDied","Data":"e022b16bf51f1e5059fe0261194348570e6ab0ade53b09ae7a127c388b5a7dc5"} Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.478271 4677 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-f5cx5" Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.478277 4677 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e022b16bf51f1e5059fe0261194348570e6ab0ade53b09ae7a127c388b5a7dc5" Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.543445 4677 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q"] Dec 03 16:00:04 crc kubenswrapper[4677]: I1203 16:00:04.554533 4677 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-nm44q"] Dec 03 16:00:05 crc kubenswrapper[4677]: I1203 16:00:05.993580 4677 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aecb6a0-009d-43bb-a67f-61e0c2978610" path="/var/lib/kubelet/pods/4aecb6a0-009d-43bb-a67f-61e0c2978610/volumes" Dec 03 16:00:08 crc kubenswrapper[4677]: I1203 16:00:08.437258 4677 patch_prober.go:28] interesting pod/machine-config-daemon-2bx7x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:00:08 crc kubenswrapper[4677]: I1203 16:00:08.438248 4677 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:00:08 crc kubenswrapper[4677]: I1203 16:00:08.438330 4677 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" Dec 03 16:00:08 crc kubenswrapper[4677]: I1203 16:00:08.439701 4677 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"573e5375928362f9817cbfab82c4bb837d24528e023e4f9d74d58886defab8fa"} pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:00:08 crc kubenswrapper[4677]: I1203 16:00:08.439786 4677 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerName="machine-config-daemon" containerID="cri-o://573e5375928362f9817cbfab82c4bb837d24528e023e4f9d74d58886defab8fa" gracePeriod=600 Dec 03 16:00:08 crc kubenswrapper[4677]: E1203 16:00:08.704914 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" Dec 03 16:00:09 crc kubenswrapper[4677]: I1203 16:00:09.532566 4677 generic.go:334] "Generic (PLEG): container finished" podID="c1665842-4668-4fed-a04f-cf4aa0043ebc" containerID="573e5375928362f9817cbfab82c4bb837d24528e023e4f9d74d58886defab8fa" exitCode=0 Dec 03 16:00:09 crc kubenswrapper[4677]: I1203 16:00:09.533250 4677 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" event={"ID":"c1665842-4668-4fed-a04f-cf4aa0043ebc","Type":"ContainerDied","Data":"573e5375928362f9817cbfab82c4bb837d24528e023e4f9d74d58886defab8fa"} Dec 03 16:00:09 crc kubenswrapper[4677]: I1203 16:00:09.533309 4677 scope.go:117] "RemoveContainer" containerID="4bf1d337bf6e33dc878f09d0f5a4457a9b07f765334ef5c2d68bc844e60b6241" Dec 03 16:00:09 crc kubenswrapper[4677]: I1203 16:00:09.534671 4677 scope.go:117] "RemoveContainer" containerID="573e5375928362f9817cbfab82c4bb837d24528e023e4f9d74d58886defab8fa" Dec 03 16:00:09 crc kubenswrapper[4677]: E1203 16:00:09.535183 4677 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2bx7x_openshift-machine-config-operator(c1665842-4668-4fed-a04f-cf4aa0043ebc)\"" pod="openshift-machine-config-operator/machine-config-daemon-2bx7x" podUID="c1665842-4668-4fed-a04f-cf4aa0043ebc" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114057232024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114057233017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114037120016477 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114037120015447 5ustar corecore